perf_event.c 29 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953954955956957958959960961962963964965966967968969970971972973974975976977978979980981982983984985986987988989990991992993994995996997998999100010011002100310041005100610071008100910101011101210131014101510161017101810191020102110221023102410251026102710281029103010311032103310341035103610371038103910401041104210431044104510461047104810491050105110521053105410551056105710581059106010611062106310641065106610671068106910701071107210731074107510761077107810791080108110821083108410851086108710881089109010911092109310941095109610971098109911001101110211031104110511061107110811091110111111121113111411151116111711181119112011211122112311241125112611271128112911301131113211331134113511361137113811391140114111421143114411451146114711481149115011511152115311541155115611571158115911601161116211631164116511661167116811691170117111721173117411751176117711781179118011811182118311841185118611871188118911901191119211931194119511961197119811991200120112021203120412051206120712081209121012111212121312141215121612171218121912201221122212231224122512261227122812291230123112321233123412351236123712381239124012411242124312441245124612471248124912501251125212531254125512561257125812591260126112621263126412651266126712681269127012711272127312741275127612771278127912801281128212831284128512861287128812891290129112921293129412951296129712981299130013011302130313041305130613071308130913101311131213131314131513161317131813191320132113221323132413251326132713281329133013311332133313341335133613371338133913401341134213431344134513461347134813491350135113521353135413551356135713581359136013611362136313641365136613671368136913701371
  1. // SPDX-License-Identifier: GPL-2.0
  2. // Copyright (C) 2018 Hangzhou C-SKY Microsystems co.,ltd.
  3. #include <linux/errno.h>
  4. #include <linux/interrupt.h>
  5. #include <linux/module.h>
  6. #include <linux/of.h>
  7. #include <linux/perf_event.h>
  8. #include <linux/platform_device.h>
  9. #define CSKY_PMU_MAX_EVENTS 32
  10. #define DEFAULT_COUNT_WIDTH 48
  11. #define HPCR "<0, 0x0>" /* PMU Control reg */
  12. #define HPSPR "<0, 0x1>" /* Start PC reg */
  13. #define HPEPR "<0, 0x2>" /* End PC reg */
  14. #define HPSIR "<0, 0x3>" /* Soft Counter reg */
  15. #define HPCNTENR "<0, 0x4>" /* Count Enable reg */
  16. #define HPINTENR "<0, 0x5>" /* Interrupt Enable reg */
  17. #define HPOFSR "<0, 0x6>" /* Interrupt Status reg */
  18. /* The events for a given PMU register set. */
  19. struct pmu_hw_events {
  20. /*
  21. * The events that are active on the PMU for the given index.
  22. */
  23. struct perf_event *events[CSKY_PMU_MAX_EVENTS];
  24. /*
  25. * A 1 bit for an index indicates that the counter is being used for
  26. * an event. A 0 means that the counter can be used.
  27. */
  28. unsigned long used_mask[BITS_TO_LONGS(CSKY_PMU_MAX_EVENTS)];
  29. };
  30. static uint64_t (*hw_raw_read_mapping[CSKY_PMU_MAX_EVENTS])(void);
  31. static void (*hw_raw_write_mapping[CSKY_PMU_MAX_EVENTS])(uint64_t val);
  32. static struct csky_pmu_t {
  33. struct pmu pmu;
  34. struct pmu_hw_events __percpu *hw_events;
  35. struct platform_device *plat_device;
  36. uint32_t count_width;
  37. uint32_t hpcr;
  38. u64 max_period;
  39. } csky_pmu;
  40. static int csky_pmu_irq;
  41. #define to_csky_pmu(p) (container_of(p, struct csky_pmu, pmu))
  42. #define cprgr(reg) \
  43. ({ \
  44. unsigned int tmp; \
  45. asm volatile("cprgr %0, "reg"\n" \
  46. : "=r"(tmp) \
  47. : \
  48. : "memory"); \
  49. tmp; \
  50. })
  51. #define cpwgr(reg, val) \
  52. ({ \
  53. asm volatile( \
  54. "cpwgr %0, "reg"\n" \
  55. : \
  56. : "r"(val) \
  57. : "memory"); \
  58. })
  59. #define cprcr(reg) \
  60. ({ \
  61. unsigned int tmp; \
  62. asm volatile("cprcr %0, "reg"\n" \
  63. : "=r"(tmp) \
  64. : \
  65. : "memory"); \
  66. tmp; \
  67. })
  68. #define cpwcr(reg, val) \
  69. ({ \
  70. asm volatile( \
  71. "cpwcr %0, "reg"\n" \
  72. : \
  73. : "r"(val) \
  74. : "memory"); \
  75. })
  76. /* cycle counter */
  77. uint64_t csky_pmu_read_cc(void)
  78. {
  79. uint32_t lo, hi, tmp;
  80. uint64_t result;
  81. do {
  82. tmp = cprgr("<0, 0x3>");
  83. lo = cprgr("<0, 0x2>");
  84. hi = cprgr("<0, 0x3>");
  85. } while (hi != tmp);
  86. result = (uint64_t) (hi) << 32;
  87. result |= lo;
  88. return result;
  89. }
  90. static void csky_pmu_write_cc(uint64_t val)
  91. {
  92. cpwgr("<0, 0x2>", (uint32_t) val);
  93. cpwgr("<0, 0x3>", (uint32_t) (val >> 32));
  94. }
  95. /* instruction counter */
  96. static uint64_t csky_pmu_read_ic(void)
  97. {
  98. uint32_t lo, hi, tmp;
  99. uint64_t result;
  100. do {
  101. tmp = cprgr("<0, 0x5>");
  102. lo = cprgr("<0, 0x4>");
  103. hi = cprgr("<0, 0x5>");
  104. } while (hi != tmp);
  105. result = (uint64_t) (hi) << 32;
  106. result |= lo;
  107. return result;
  108. }
  109. static void csky_pmu_write_ic(uint64_t val)
  110. {
  111. cpwgr("<0, 0x4>", (uint32_t) val);
  112. cpwgr("<0, 0x5>", (uint32_t) (val >> 32));
  113. }
  114. /* l1 icache access counter */
  115. static uint64_t csky_pmu_read_icac(void)
  116. {
  117. uint32_t lo, hi, tmp;
  118. uint64_t result;
  119. do {
  120. tmp = cprgr("<0, 0x7>");
  121. lo = cprgr("<0, 0x6>");
  122. hi = cprgr("<0, 0x7>");
  123. } while (hi != tmp);
  124. result = (uint64_t) (hi) << 32;
  125. result |= lo;
  126. return result;
  127. }
  128. static void csky_pmu_write_icac(uint64_t val)
  129. {
  130. cpwgr("<0, 0x6>", (uint32_t) val);
  131. cpwgr("<0, 0x7>", (uint32_t) (val >> 32));
  132. }
  133. /* l1 icache miss counter */
  134. static uint64_t csky_pmu_read_icmc(void)
  135. {
  136. uint32_t lo, hi, tmp;
  137. uint64_t result;
  138. do {
  139. tmp = cprgr("<0, 0x9>");
  140. lo = cprgr("<0, 0x8>");
  141. hi = cprgr("<0, 0x9>");
  142. } while (hi != tmp);
  143. result = (uint64_t) (hi) << 32;
  144. result |= lo;
  145. return result;
  146. }
  147. static void csky_pmu_write_icmc(uint64_t val)
  148. {
  149. cpwgr("<0, 0x8>", (uint32_t) val);
  150. cpwgr("<0, 0x9>", (uint32_t) (val >> 32));
  151. }
  152. /* l1 dcache access counter */
  153. static uint64_t csky_pmu_read_dcac(void)
  154. {
  155. uint32_t lo, hi, tmp;
  156. uint64_t result;
  157. do {
  158. tmp = cprgr("<0, 0xb>");
  159. lo = cprgr("<0, 0xa>");
  160. hi = cprgr("<0, 0xb>");
  161. } while (hi != tmp);
  162. result = (uint64_t) (hi) << 32;
  163. result |= lo;
  164. return result;
  165. }
  166. static void csky_pmu_write_dcac(uint64_t val)
  167. {
  168. cpwgr("<0, 0xa>", (uint32_t) val);
  169. cpwgr("<0, 0xb>", (uint32_t) (val >> 32));
  170. }
  171. /* l1 dcache miss counter */
  172. static uint64_t csky_pmu_read_dcmc(void)
  173. {
  174. uint32_t lo, hi, tmp;
  175. uint64_t result;
  176. do {
  177. tmp = cprgr("<0, 0xd>");
  178. lo = cprgr("<0, 0xc>");
  179. hi = cprgr("<0, 0xd>");
  180. } while (hi != tmp);
  181. result = (uint64_t) (hi) << 32;
  182. result |= lo;
  183. return result;
  184. }
  185. static void csky_pmu_write_dcmc(uint64_t val)
  186. {
  187. cpwgr("<0, 0xc>", (uint32_t) val);
  188. cpwgr("<0, 0xd>", (uint32_t) (val >> 32));
  189. }
  190. /* l2 cache access counter */
  191. static uint64_t csky_pmu_read_l2ac(void)
  192. {
  193. uint32_t lo, hi, tmp;
  194. uint64_t result;
  195. do {
  196. tmp = cprgr("<0, 0xf>");
  197. lo = cprgr("<0, 0xe>");
  198. hi = cprgr("<0, 0xf>");
  199. } while (hi != tmp);
  200. result = (uint64_t) (hi) << 32;
  201. result |= lo;
  202. return result;
  203. }
  204. static void csky_pmu_write_l2ac(uint64_t val)
  205. {
  206. cpwgr("<0, 0xe>", (uint32_t) val);
  207. cpwgr("<0, 0xf>", (uint32_t) (val >> 32));
  208. }
  209. /* l2 cache miss counter */
  210. static uint64_t csky_pmu_read_l2mc(void)
  211. {
  212. uint32_t lo, hi, tmp;
  213. uint64_t result;
  214. do {
  215. tmp = cprgr("<0, 0x11>");
  216. lo = cprgr("<0, 0x10>");
  217. hi = cprgr("<0, 0x11>");
  218. } while (hi != tmp);
  219. result = (uint64_t) (hi) << 32;
  220. result |= lo;
  221. return result;
  222. }
  223. static void csky_pmu_write_l2mc(uint64_t val)
  224. {
  225. cpwgr("<0, 0x10>", (uint32_t) val);
  226. cpwgr("<0, 0x11>", (uint32_t) (val >> 32));
  227. }
  228. /* I-UTLB miss counter */
  229. static uint64_t csky_pmu_read_iutlbmc(void)
  230. {
  231. uint32_t lo, hi, tmp;
  232. uint64_t result;
  233. do {
  234. tmp = cprgr("<0, 0x15>");
  235. lo = cprgr("<0, 0x14>");
  236. hi = cprgr("<0, 0x15>");
  237. } while (hi != tmp);
  238. result = (uint64_t) (hi) << 32;
  239. result |= lo;
  240. return result;
  241. }
  242. static void csky_pmu_write_iutlbmc(uint64_t val)
  243. {
  244. cpwgr("<0, 0x14>", (uint32_t) val);
  245. cpwgr("<0, 0x15>", (uint32_t) (val >> 32));
  246. }
  247. /* D-UTLB miss counter */
  248. static uint64_t csky_pmu_read_dutlbmc(void)
  249. {
  250. uint32_t lo, hi, tmp;
  251. uint64_t result;
  252. do {
  253. tmp = cprgr("<0, 0x17>");
  254. lo = cprgr("<0, 0x16>");
  255. hi = cprgr("<0, 0x17>");
  256. } while (hi != tmp);
  257. result = (uint64_t) (hi) << 32;
  258. result |= lo;
  259. return result;
  260. }
  261. static void csky_pmu_write_dutlbmc(uint64_t val)
  262. {
  263. cpwgr("<0, 0x16>", (uint32_t) val);
  264. cpwgr("<0, 0x17>", (uint32_t) (val >> 32));
  265. }
  266. /* JTLB miss counter */
  267. static uint64_t csky_pmu_read_jtlbmc(void)
  268. {
  269. uint32_t lo, hi, tmp;
  270. uint64_t result;
  271. do {
  272. tmp = cprgr("<0, 0x19>");
  273. lo = cprgr("<0, 0x18>");
  274. hi = cprgr("<0, 0x19>");
  275. } while (hi != tmp);
  276. result = (uint64_t) (hi) << 32;
  277. result |= lo;
  278. return result;
  279. }
  280. static void csky_pmu_write_jtlbmc(uint64_t val)
  281. {
  282. cpwgr("<0, 0x18>", (uint32_t) val);
  283. cpwgr("<0, 0x19>", (uint32_t) (val >> 32));
  284. }
  285. /* software counter */
  286. static uint64_t csky_pmu_read_softc(void)
  287. {
  288. uint32_t lo, hi, tmp;
  289. uint64_t result;
  290. do {
  291. tmp = cprgr("<0, 0x1b>");
  292. lo = cprgr("<0, 0x1a>");
  293. hi = cprgr("<0, 0x1b>");
  294. } while (hi != tmp);
  295. result = (uint64_t) (hi) << 32;
  296. result |= lo;
  297. return result;
  298. }
  299. static void csky_pmu_write_softc(uint64_t val)
  300. {
  301. cpwgr("<0, 0x1a>", (uint32_t) val);
  302. cpwgr("<0, 0x1b>", (uint32_t) (val >> 32));
  303. }
  304. /* conditional branch mispredict counter */
  305. static uint64_t csky_pmu_read_cbmc(void)
  306. {
  307. uint32_t lo, hi, tmp;
  308. uint64_t result;
  309. do {
  310. tmp = cprgr("<0, 0x1d>");
  311. lo = cprgr("<0, 0x1c>");
  312. hi = cprgr("<0, 0x1d>");
  313. } while (hi != tmp);
  314. result = (uint64_t) (hi) << 32;
  315. result |= lo;
  316. return result;
  317. }
  318. static void csky_pmu_write_cbmc(uint64_t val)
  319. {
  320. cpwgr("<0, 0x1c>", (uint32_t) val);
  321. cpwgr("<0, 0x1d>", (uint32_t) (val >> 32));
  322. }
  323. /* conditional branch instruction counter */
  324. static uint64_t csky_pmu_read_cbic(void)
  325. {
  326. uint32_t lo, hi, tmp;
  327. uint64_t result;
  328. do {
  329. tmp = cprgr("<0, 0x1f>");
  330. lo = cprgr("<0, 0x1e>");
  331. hi = cprgr("<0, 0x1f>");
  332. } while (hi != tmp);
  333. result = (uint64_t) (hi) << 32;
  334. result |= lo;
  335. return result;
  336. }
  337. static void csky_pmu_write_cbic(uint64_t val)
  338. {
  339. cpwgr("<0, 0x1e>", (uint32_t) val);
  340. cpwgr("<0, 0x1f>", (uint32_t) (val >> 32));
  341. }
  342. /* indirect branch mispredict counter */
  343. static uint64_t csky_pmu_read_ibmc(void)
  344. {
  345. uint32_t lo, hi, tmp;
  346. uint64_t result;
  347. do {
  348. tmp = cprgr("<0, 0x21>");
  349. lo = cprgr("<0, 0x20>");
  350. hi = cprgr("<0, 0x21>");
  351. } while (hi != tmp);
  352. result = (uint64_t) (hi) << 32;
  353. result |= lo;
  354. return result;
  355. }
  356. static void csky_pmu_write_ibmc(uint64_t val)
  357. {
  358. cpwgr("<0, 0x20>", (uint32_t) val);
  359. cpwgr("<0, 0x21>", (uint32_t) (val >> 32));
  360. }
  361. /* indirect branch instruction counter */
  362. static uint64_t csky_pmu_read_ibic(void)
  363. {
  364. uint32_t lo, hi, tmp;
  365. uint64_t result;
  366. do {
  367. tmp = cprgr("<0, 0x23>");
  368. lo = cprgr("<0, 0x22>");
  369. hi = cprgr("<0, 0x23>");
  370. } while (hi != tmp);
  371. result = (uint64_t) (hi) << 32;
  372. result |= lo;
  373. return result;
  374. }
  375. static void csky_pmu_write_ibic(uint64_t val)
  376. {
  377. cpwgr("<0, 0x22>", (uint32_t) val);
  378. cpwgr("<0, 0x23>", (uint32_t) (val >> 32));
  379. }
  380. /* LSU spec fail counter */
  381. static uint64_t csky_pmu_read_lsfc(void)
  382. {
  383. uint32_t lo, hi, tmp;
  384. uint64_t result;
  385. do {
  386. tmp = cprgr("<0, 0x25>");
  387. lo = cprgr("<0, 0x24>");
  388. hi = cprgr("<0, 0x25>");
  389. } while (hi != tmp);
  390. result = (uint64_t) (hi) << 32;
  391. result |= lo;
  392. return result;
  393. }
  394. static void csky_pmu_write_lsfc(uint64_t val)
  395. {
  396. cpwgr("<0, 0x24>", (uint32_t) val);
  397. cpwgr("<0, 0x25>", (uint32_t) (val >> 32));
  398. }
  399. /* store instruction counter */
  400. static uint64_t csky_pmu_read_sic(void)
  401. {
  402. uint32_t lo, hi, tmp;
  403. uint64_t result;
  404. do {
  405. tmp = cprgr("<0, 0x27>");
  406. lo = cprgr("<0, 0x26>");
  407. hi = cprgr("<0, 0x27>");
  408. } while (hi != tmp);
  409. result = (uint64_t) (hi) << 32;
  410. result |= lo;
  411. return result;
  412. }
  413. static void csky_pmu_write_sic(uint64_t val)
  414. {
  415. cpwgr("<0, 0x26>", (uint32_t) val);
  416. cpwgr("<0, 0x27>", (uint32_t) (val >> 32));
  417. }
  418. /* dcache read access counter */
  419. static uint64_t csky_pmu_read_dcrac(void)
  420. {
  421. uint32_t lo, hi, tmp;
  422. uint64_t result;
  423. do {
  424. tmp = cprgr("<0, 0x29>");
  425. lo = cprgr("<0, 0x28>");
  426. hi = cprgr("<0, 0x29>");
  427. } while (hi != tmp);
  428. result = (uint64_t) (hi) << 32;
  429. result |= lo;
  430. return result;
  431. }
  432. static void csky_pmu_write_dcrac(uint64_t val)
  433. {
  434. cpwgr("<0, 0x28>", (uint32_t) val);
  435. cpwgr("<0, 0x29>", (uint32_t) (val >> 32));
  436. }
  437. /* dcache read miss counter */
  438. static uint64_t csky_pmu_read_dcrmc(void)
  439. {
  440. uint32_t lo, hi, tmp;
  441. uint64_t result;
  442. do {
  443. tmp = cprgr("<0, 0x2b>");
  444. lo = cprgr("<0, 0x2a>");
  445. hi = cprgr("<0, 0x2b>");
  446. } while (hi != tmp);
  447. result = (uint64_t) (hi) << 32;
  448. result |= lo;
  449. return result;
  450. }
  451. static void csky_pmu_write_dcrmc(uint64_t val)
  452. {
  453. cpwgr("<0, 0x2a>", (uint32_t) val);
  454. cpwgr("<0, 0x2b>", (uint32_t) (val >> 32));
  455. }
  456. /* dcache write access counter */
  457. static uint64_t csky_pmu_read_dcwac(void)
  458. {
  459. uint32_t lo, hi, tmp;
  460. uint64_t result;
  461. do {
  462. tmp = cprgr("<0, 0x2d>");
  463. lo = cprgr("<0, 0x2c>");
  464. hi = cprgr("<0, 0x2d>");
  465. } while (hi != tmp);
  466. result = (uint64_t) (hi) << 32;
  467. result |= lo;
  468. return result;
  469. }
  470. static void csky_pmu_write_dcwac(uint64_t val)
  471. {
  472. cpwgr("<0, 0x2c>", (uint32_t) val);
  473. cpwgr("<0, 0x2d>", (uint32_t) (val >> 32));
  474. }
  475. /* dcache write miss counter */
  476. static uint64_t csky_pmu_read_dcwmc(void)
  477. {
  478. uint32_t lo, hi, tmp;
  479. uint64_t result;
  480. do {
  481. tmp = cprgr("<0, 0x2f>");
  482. lo = cprgr("<0, 0x2e>");
  483. hi = cprgr("<0, 0x2f>");
  484. } while (hi != tmp);
  485. result = (uint64_t) (hi) << 32;
  486. result |= lo;
  487. return result;
  488. }
  489. static void csky_pmu_write_dcwmc(uint64_t val)
  490. {
  491. cpwgr("<0, 0x2e>", (uint32_t) val);
  492. cpwgr("<0, 0x2f>", (uint32_t) (val >> 32));
  493. }
  494. /* l2cache read access counter */
  495. static uint64_t csky_pmu_read_l2rac(void)
  496. {
  497. uint32_t lo, hi, tmp;
  498. uint64_t result;
  499. do {
  500. tmp = cprgr("<0, 0x31>");
  501. lo = cprgr("<0, 0x30>");
  502. hi = cprgr("<0, 0x31>");
  503. } while (hi != tmp);
  504. result = (uint64_t) (hi) << 32;
  505. result |= lo;
  506. return result;
  507. }
  508. static void csky_pmu_write_l2rac(uint64_t val)
  509. {
  510. cpwgr("<0, 0x30>", (uint32_t) val);
  511. cpwgr("<0, 0x31>", (uint32_t) (val >> 32));
  512. }
  513. /* l2cache read miss counter */
  514. static uint64_t csky_pmu_read_l2rmc(void)
  515. {
  516. uint32_t lo, hi, tmp;
  517. uint64_t result;
  518. do {
  519. tmp = cprgr("<0, 0x33>");
  520. lo = cprgr("<0, 0x32>");
  521. hi = cprgr("<0, 0x33>");
  522. } while (hi != tmp);
  523. result = (uint64_t) (hi) << 32;
  524. result |= lo;
  525. return result;
  526. }
  527. static void csky_pmu_write_l2rmc(uint64_t val)
  528. {
  529. cpwgr("<0, 0x32>", (uint32_t) val);
  530. cpwgr("<0, 0x33>", (uint32_t) (val >> 32));
  531. }
  532. /* l2cache write access counter */
  533. static uint64_t csky_pmu_read_l2wac(void)
  534. {
  535. uint32_t lo, hi, tmp;
  536. uint64_t result;
  537. do {
  538. tmp = cprgr("<0, 0x35>");
  539. lo = cprgr("<0, 0x34>");
  540. hi = cprgr("<0, 0x35>");
  541. } while (hi != tmp);
  542. result = (uint64_t) (hi) << 32;
  543. result |= lo;
  544. return result;
  545. }
  546. static void csky_pmu_write_l2wac(uint64_t val)
  547. {
  548. cpwgr("<0, 0x34>", (uint32_t) val);
  549. cpwgr("<0, 0x35>", (uint32_t) (val >> 32));
  550. }
  551. /* l2cache write miss counter */
  552. static uint64_t csky_pmu_read_l2wmc(void)
  553. {
  554. uint32_t lo, hi, tmp;
  555. uint64_t result;
  556. do {
  557. tmp = cprgr("<0, 0x37>");
  558. lo = cprgr("<0, 0x36>");
  559. hi = cprgr("<0, 0x37>");
  560. } while (hi != tmp);
  561. result = (uint64_t) (hi) << 32;
  562. result |= lo;
  563. return result;
  564. }
  565. static void csky_pmu_write_l2wmc(uint64_t val)
  566. {
  567. cpwgr("<0, 0x36>", (uint32_t) val);
  568. cpwgr("<0, 0x37>", (uint32_t) (val >> 32));
  569. }
  570. #define HW_OP_UNSUPPORTED 0xffff
  571. static const int csky_pmu_hw_map[PERF_COUNT_HW_MAX] = {
  572. [PERF_COUNT_HW_CPU_CYCLES] = 0x1,
  573. [PERF_COUNT_HW_INSTRUCTIONS] = 0x2,
  574. [PERF_COUNT_HW_CACHE_REFERENCES] = HW_OP_UNSUPPORTED,
  575. [PERF_COUNT_HW_CACHE_MISSES] = HW_OP_UNSUPPORTED,
  576. [PERF_COUNT_HW_BRANCH_INSTRUCTIONS] = 0xf,
  577. [PERF_COUNT_HW_BRANCH_MISSES] = 0xe,
  578. [PERF_COUNT_HW_BUS_CYCLES] = HW_OP_UNSUPPORTED,
  579. [PERF_COUNT_HW_STALLED_CYCLES_FRONTEND] = HW_OP_UNSUPPORTED,
  580. [PERF_COUNT_HW_STALLED_CYCLES_BACKEND] = HW_OP_UNSUPPORTED,
  581. [PERF_COUNT_HW_REF_CPU_CYCLES] = HW_OP_UNSUPPORTED,
  582. };
  583. #define C(_x) PERF_COUNT_HW_CACHE_##_x
  584. #define CACHE_OP_UNSUPPORTED 0xffff
  585. static const int csky_pmu_cache_map[C(MAX)][C(OP_MAX)][C(RESULT_MAX)] = {
  586. [C(L1D)] = {
  587. #ifdef CONFIG_CPU_CK810
  588. [C(OP_READ)] = {
  589. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  590. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  591. },
  592. [C(OP_WRITE)] = {
  593. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  594. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  595. },
  596. [C(OP_PREFETCH)] = {
  597. [C(RESULT_ACCESS)] = 0x5,
  598. [C(RESULT_MISS)] = 0x6,
  599. },
  600. #else
  601. [C(OP_READ)] = {
  602. [C(RESULT_ACCESS)] = 0x14,
  603. [C(RESULT_MISS)] = 0x15,
  604. },
  605. [C(OP_WRITE)] = {
  606. [C(RESULT_ACCESS)] = 0x16,
  607. [C(RESULT_MISS)] = 0x17,
  608. },
  609. [C(OP_PREFETCH)] = {
  610. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  611. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  612. },
  613. #endif
  614. },
  615. [C(L1I)] = {
  616. [C(OP_READ)] = {
  617. [C(RESULT_ACCESS)] = 0x3,
  618. [C(RESULT_MISS)] = 0x4,
  619. },
  620. [C(OP_WRITE)] = {
  621. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  622. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  623. },
  624. [C(OP_PREFETCH)] = {
  625. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  626. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  627. },
  628. },
  629. [C(LL)] = {
  630. #ifdef CONFIG_CPU_CK810
  631. [C(OP_READ)] = {
  632. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  633. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  634. },
  635. [C(OP_WRITE)] = {
  636. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  637. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  638. },
  639. [C(OP_PREFETCH)] = {
  640. [C(RESULT_ACCESS)] = 0x7,
  641. [C(RESULT_MISS)] = 0x8,
  642. },
  643. #else
  644. [C(OP_READ)] = {
  645. [C(RESULT_ACCESS)] = 0x18,
  646. [C(RESULT_MISS)] = 0x19,
  647. },
  648. [C(OP_WRITE)] = {
  649. [C(RESULT_ACCESS)] = 0x1a,
  650. [C(RESULT_MISS)] = 0x1b,
  651. },
  652. [C(OP_PREFETCH)] = {
  653. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  654. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  655. },
  656. #endif
  657. },
  658. [C(DTLB)] = {
  659. #ifdef CONFIG_CPU_CK810
  660. [C(OP_READ)] = {
  661. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  662. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  663. },
  664. [C(OP_WRITE)] = {
  665. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  666. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  667. },
  668. #else
  669. [C(OP_READ)] = {
  670. [C(RESULT_ACCESS)] = 0x14,
  671. [C(RESULT_MISS)] = 0xb,
  672. },
  673. [C(OP_WRITE)] = {
  674. [C(RESULT_ACCESS)] = 0x16,
  675. [C(RESULT_MISS)] = 0xb,
  676. },
  677. #endif
  678. [C(OP_PREFETCH)] = {
  679. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  680. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  681. },
  682. },
  683. [C(ITLB)] = {
  684. #ifdef CONFIG_CPU_CK810
  685. [C(OP_READ)] = {
  686. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  687. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  688. },
  689. #else
  690. [C(OP_READ)] = {
  691. [C(RESULT_ACCESS)] = 0x3,
  692. [C(RESULT_MISS)] = 0xa,
  693. },
  694. #endif
  695. [C(OP_WRITE)] = {
  696. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  697. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  698. },
  699. [C(OP_PREFETCH)] = {
  700. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  701. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  702. },
  703. },
  704. [C(BPU)] = {
  705. [C(OP_READ)] = {
  706. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  707. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  708. },
  709. [C(OP_WRITE)] = {
  710. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  711. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  712. },
  713. [C(OP_PREFETCH)] = {
  714. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  715. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  716. },
  717. },
  718. [C(NODE)] = {
  719. [C(OP_READ)] = {
  720. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  721. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  722. },
  723. [C(OP_WRITE)] = {
  724. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  725. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  726. },
  727. [C(OP_PREFETCH)] = {
  728. [C(RESULT_ACCESS)] = CACHE_OP_UNSUPPORTED,
  729. [C(RESULT_MISS)] = CACHE_OP_UNSUPPORTED,
  730. },
  731. },
  732. };
  733. int csky_pmu_event_set_period(struct perf_event *event)
  734. {
  735. struct hw_perf_event *hwc = &event->hw;
  736. s64 left = local64_read(&hwc->period_left);
  737. s64 period = hwc->sample_period;
  738. int ret = 0;
  739. if (unlikely(left <= -period)) {
  740. left = period;
  741. local64_set(&hwc->period_left, left);
  742. hwc->last_period = period;
  743. ret = 1;
  744. }
  745. if (unlikely(left <= 0)) {
  746. left += period;
  747. local64_set(&hwc->period_left, left);
  748. hwc->last_period = period;
  749. ret = 1;
  750. }
  751. if (left > (s64)csky_pmu.max_period)
  752. left = csky_pmu.max_period;
  753. /*
  754. * The hw event starts counting from this event offset,
  755. * mark it to be able to extract future "deltas":
  756. */
  757. local64_set(&hwc->prev_count, (u64)(-left));
  758. if (hw_raw_write_mapping[hwc->idx] != NULL)
  759. hw_raw_write_mapping[hwc->idx]((u64)(-left) &
  760. csky_pmu.max_period);
  761. cpwcr(HPOFSR, ~BIT(hwc->idx) & cprcr(HPOFSR));
  762. perf_event_update_userpage(event);
  763. return ret;
  764. }
  765. static void csky_perf_event_update(struct perf_event *event,
  766. struct hw_perf_event *hwc)
  767. {
  768. uint64_t prev_raw_count = local64_read(&hwc->prev_count);
  769. /*
  770. * Sign extend count value to 64bit, otherwise delta calculation
  771. * would be incorrect when overflow occurs.
  772. */
  773. uint64_t new_raw_count = sign_extend64(
  774. hw_raw_read_mapping[hwc->idx](), csky_pmu.count_width - 1);
  775. int64_t delta = new_raw_count - prev_raw_count;
  776. /*
  777. * We aren't afraid of hwc->prev_count changing beneath our feet
  778. * because there's no way for us to re-enter this function anytime.
  779. */
  780. local64_set(&hwc->prev_count, new_raw_count);
  781. local64_add(delta, &event->count);
  782. local64_sub(delta, &hwc->period_left);
  783. }
  784. static void csky_pmu_reset(void *info)
  785. {
  786. cpwcr(HPCR, BIT(31) | BIT(30) | BIT(1));
  787. }
  788. static void csky_pmu_read(struct perf_event *event)
  789. {
  790. csky_perf_event_update(event, &event->hw);
  791. }
  792. static int csky_pmu_cache_event(u64 config)
  793. {
  794. unsigned int cache_type, cache_op, cache_result;
  795. cache_type = (config >> 0) & 0xff;
  796. cache_op = (config >> 8) & 0xff;
  797. cache_result = (config >> 16) & 0xff;
  798. if (cache_type >= PERF_COUNT_HW_CACHE_MAX)
  799. return -EINVAL;
  800. if (cache_op >= PERF_COUNT_HW_CACHE_OP_MAX)
  801. return -EINVAL;
  802. if (cache_result >= PERF_COUNT_HW_CACHE_RESULT_MAX)
  803. return -EINVAL;
  804. return csky_pmu_cache_map[cache_type][cache_op][cache_result];
  805. }
  806. static int csky_pmu_event_init(struct perf_event *event)
  807. {
  808. struct hw_perf_event *hwc = &event->hw;
  809. int ret;
  810. switch (event->attr.type) {
  811. case PERF_TYPE_HARDWARE:
  812. if (event->attr.config >= PERF_COUNT_HW_MAX)
  813. return -ENOENT;
  814. ret = csky_pmu_hw_map[event->attr.config];
  815. if (ret == HW_OP_UNSUPPORTED)
  816. return -ENOENT;
  817. hwc->idx = ret;
  818. break;
  819. case PERF_TYPE_HW_CACHE:
  820. ret = csky_pmu_cache_event(event->attr.config);
  821. if (ret == CACHE_OP_UNSUPPORTED)
  822. return -ENOENT;
  823. hwc->idx = ret;
  824. break;
  825. case PERF_TYPE_RAW:
  826. if (hw_raw_read_mapping[event->attr.config] == NULL)
  827. return -ENOENT;
  828. hwc->idx = event->attr.config;
  829. break;
  830. default:
  831. return -ENOENT;
  832. }
  833. if (event->attr.exclude_user)
  834. csky_pmu.hpcr = BIT(2);
  835. else if (event->attr.exclude_kernel)
  836. csky_pmu.hpcr = BIT(3);
  837. else
  838. csky_pmu.hpcr = BIT(2) | BIT(3);
  839. csky_pmu.hpcr |= BIT(1) | BIT(0);
  840. return 0;
  841. }
  842. /* starts all counters */
  843. static void csky_pmu_enable(struct pmu *pmu)
  844. {
  845. cpwcr(HPCR, csky_pmu.hpcr);
  846. }
  847. /* stops all counters */
  848. static void csky_pmu_disable(struct pmu *pmu)
  849. {
  850. cpwcr(HPCR, BIT(1));
  851. }
  852. static void csky_pmu_start(struct perf_event *event, int flags)
  853. {
  854. unsigned long flg;
  855. struct hw_perf_event *hwc = &event->hw;
  856. int idx = hwc->idx;
  857. if (WARN_ON_ONCE(idx == -1))
  858. return;
  859. if (flags & PERF_EF_RELOAD)
  860. WARN_ON_ONCE(!(hwc->state & PERF_HES_UPTODATE));
  861. hwc->state = 0;
  862. csky_pmu_event_set_period(event);
  863. local_irq_save(flg);
  864. cpwcr(HPINTENR, BIT(idx) | cprcr(HPINTENR));
  865. cpwcr(HPCNTENR, BIT(idx) | cprcr(HPCNTENR));
  866. local_irq_restore(flg);
  867. }
  868. static void csky_pmu_stop_event(struct perf_event *event)
  869. {
  870. unsigned long flg;
  871. struct hw_perf_event *hwc = &event->hw;
  872. int idx = hwc->idx;
  873. local_irq_save(flg);
  874. cpwcr(HPINTENR, ~BIT(idx) & cprcr(HPINTENR));
  875. cpwcr(HPCNTENR, ~BIT(idx) & cprcr(HPCNTENR));
  876. local_irq_restore(flg);
  877. }
  878. static void csky_pmu_stop(struct perf_event *event, int flags)
  879. {
  880. if (!(event->hw.state & PERF_HES_STOPPED)) {
  881. csky_pmu_stop_event(event);
  882. event->hw.state |= PERF_HES_STOPPED;
  883. }
  884. if ((flags & PERF_EF_UPDATE) &&
  885. !(event->hw.state & PERF_HES_UPTODATE)) {
  886. csky_perf_event_update(event, &event->hw);
  887. event->hw.state |= PERF_HES_UPTODATE;
  888. }
  889. }
  890. static void csky_pmu_del(struct perf_event *event, int flags)
  891. {
  892. struct pmu_hw_events *hw_events = this_cpu_ptr(csky_pmu.hw_events);
  893. struct hw_perf_event *hwc = &event->hw;
  894. csky_pmu_stop(event, PERF_EF_UPDATE);
  895. hw_events->events[hwc->idx] = NULL;
  896. perf_event_update_userpage(event);
  897. }
  898. /* allocate hardware counter and optionally start counting */
  899. static int csky_pmu_add(struct perf_event *event, int flags)
  900. {
  901. struct pmu_hw_events *hw_events = this_cpu_ptr(csky_pmu.hw_events);
  902. struct hw_perf_event *hwc = &event->hw;
  903. hw_events->events[hwc->idx] = event;
  904. hwc->state = PERF_HES_UPTODATE | PERF_HES_STOPPED;
  905. if (flags & PERF_EF_START)
  906. csky_pmu_start(event, PERF_EF_RELOAD);
  907. perf_event_update_userpage(event);
  908. return 0;
  909. }
  910. static irqreturn_t csky_pmu_handle_irq(int irq_num, void *dev)
  911. {
  912. struct perf_sample_data data;
  913. struct pmu_hw_events *cpuc = this_cpu_ptr(csky_pmu.hw_events);
  914. struct pt_regs *regs;
  915. int idx;
  916. /*
  917. * Did an overflow occur?
  918. */
  919. if (!cprcr(HPOFSR))
  920. return IRQ_NONE;
  921. /*
  922. * Handle the counter(s) overflow(s)
  923. */
  924. regs = get_irq_regs();
  925. csky_pmu_disable(&csky_pmu.pmu);
  926. for (idx = 0; idx < CSKY_PMU_MAX_EVENTS; ++idx) {
  927. struct perf_event *event = cpuc->events[idx];
  928. struct hw_perf_event *hwc;
  929. /* Ignore if we don't have an event. */
  930. if (!event)
  931. continue;
  932. /*
  933. * We have a single interrupt for all counters. Check that
  934. * each counter has overflowed before we process it.
  935. */
  936. if (!(cprcr(HPOFSR) & BIT(idx)))
  937. continue;
  938. hwc = &event->hw;
  939. csky_perf_event_update(event, &event->hw);
  940. perf_sample_data_init(&data, 0, hwc->last_period);
  941. csky_pmu_event_set_period(event);
  942. if (perf_event_overflow(event, &data, regs))
  943. csky_pmu_stop_event(event);
  944. }
  945. csky_pmu_enable(&csky_pmu.pmu);
  946. /*
  947. * Handle the pending perf events.
  948. *
  949. * Note: this call *must* be run with interrupts disabled. For
  950. * platforms that can have the PMU interrupts raised as an NMI, this
  951. * will not work.
  952. */
  953. irq_work_run();
  954. return IRQ_HANDLED;
  955. }
  956. static int csky_pmu_request_irq(irq_handler_t handler)
  957. {
  958. int err, irqs;
  959. struct platform_device *pmu_device = csky_pmu.plat_device;
  960. if (!pmu_device)
  961. return -ENODEV;
  962. irqs = min(pmu_device->num_resources, num_possible_cpus());
  963. if (irqs < 1) {
  964. pr_err("no irqs for PMUs defined\n");
  965. return -ENODEV;
  966. }
  967. csky_pmu_irq = platform_get_irq(pmu_device, 0);
  968. if (csky_pmu_irq < 0)
  969. return -ENODEV;
  970. err = request_percpu_irq(csky_pmu_irq, handler, "csky-pmu",
  971. this_cpu_ptr(csky_pmu.hw_events));
  972. if (err) {
  973. pr_err("unable to request IRQ%d for CSKY PMU counters\n",
  974. csky_pmu_irq);
  975. return err;
  976. }
  977. return 0;
  978. }
  979. static void csky_pmu_free_irq(void)
  980. {
  981. int irq;
  982. struct platform_device *pmu_device = csky_pmu.plat_device;
  983. irq = platform_get_irq(pmu_device, 0);
  984. if (irq >= 0)
  985. free_percpu_irq(irq, this_cpu_ptr(csky_pmu.hw_events));
  986. }
  987. int init_hw_perf_events(void)
  988. {
  989. csky_pmu.hw_events = alloc_percpu_gfp(struct pmu_hw_events,
  990. GFP_KERNEL);
  991. if (!csky_pmu.hw_events) {
  992. pr_info("failed to allocate per-cpu PMU data.\n");
  993. return -ENOMEM;
  994. }
  995. csky_pmu.pmu = (struct pmu) {
  996. .pmu_enable = csky_pmu_enable,
  997. .pmu_disable = csky_pmu_disable,
  998. .event_init = csky_pmu_event_init,
  999. .add = csky_pmu_add,
  1000. .del = csky_pmu_del,
  1001. .start = csky_pmu_start,
  1002. .stop = csky_pmu_stop,
  1003. .read = csky_pmu_read,
  1004. };
  1005. memset((void *)hw_raw_read_mapping, 0,
  1006. sizeof(hw_raw_read_mapping[CSKY_PMU_MAX_EVENTS]));
  1007. hw_raw_read_mapping[0x1] = csky_pmu_read_cc;
  1008. hw_raw_read_mapping[0x2] = csky_pmu_read_ic;
  1009. hw_raw_read_mapping[0x3] = csky_pmu_read_icac;
  1010. hw_raw_read_mapping[0x4] = csky_pmu_read_icmc;
  1011. hw_raw_read_mapping[0x5] = csky_pmu_read_dcac;
  1012. hw_raw_read_mapping[0x6] = csky_pmu_read_dcmc;
  1013. hw_raw_read_mapping[0x7] = csky_pmu_read_l2ac;
  1014. hw_raw_read_mapping[0x8] = csky_pmu_read_l2mc;
  1015. hw_raw_read_mapping[0xa] = csky_pmu_read_iutlbmc;
  1016. hw_raw_read_mapping[0xb] = csky_pmu_read_dutlbmc;
  1017. hw_raw_read_mapping[0xc] = csky_pmu_read_jtlbmc;
  1018. hw_raw_read_mapping[0xd] = csky_pmu_read_softc;
  1019. hw_raw_read_mapping[0xe] = csky_pmu_read_cbmc;
  1020. hw_raw_read_mapping[0xf] = csky_pmu_read_cbic;
  1021. hw_raw_read_mapping[0x10] = csky_pmu_read_ibmc;
  1022. hw_raw_read_mapping[0x11] = csky_pmu_read_ibic;
  1023. hw_raw_read_mapping[0x12] = csky_pmu_read_lsfc;
  1024. hw_raw_read_mapping[0x13] = csky_pmu_read_sic;
  1025. hw_raw_read_mapping[0x14] = csky_pmu_read_dcrac;
  1026. hw_raw_read_mapping[0x15] = csky_pmu_read_dcrmc;
  1027. hw_raw_read_mapping[0x16] = csky_pmu_read_dcwac;
  1028. hw_raw_read_mapping[0x17] = csky_pmu_read_dcwmc;
  1029. hw_raw_read_mapping[0x18] = csky_pmu_read_l2rac;
  1030. hw_raw_read_mapping[0x19] = csky_pmu_read_l2rmc;
  1031. hw_raw_read_mapping[0x1a] = csky_pmu_read_l2wac;
  1032. hw_raw_read_mapping[0x1b] = csky_pmu_read_l2wmc;
  1033. memset((void *)hw_raw_write_mapping, 0,
  1034. sizeof(hw_raw_write_mapping[CSKY_PMU_MAX_EVENTS]));
  1035. hw_raw_write_mapping[0x1] = csky_pmu_write_cc;
  1036. hw_raw_write_mapping[0x2] = csky_pmu_write_ic;
  1037. hw_raw_write_mapping[0x3] = csky_pmu_write_icac;
  1038. hw_raw_write_mapping[0x4] = csky_pmu_write_icmc;
  1039. hw_raw_write_mapping[0x5] = csky_pmu_write_dcac;
  1040. hw_raw_write_mapping[0x6] = csky_pmu_write_dcmc;
  1041. hw_raw_write_mapping[0x7] = csky_pmu_write_l2ac;
  1042. hw_raw_write_mapping[0x8] = csky_pmu_write_l2mc;
  1043. hw_raw_write_mapping[0xa] = csky_pmu_write_iutlbmc;
  1044. hw_raw_write_mapping[0xb] = csky_pmu_write_dutlbmc;
  1045. hw_raw_write_mapping[0xc] = csky_pmu_write_jtlbmc;
  1046. hw_raw_write_mapping[0xd] = csky_pmu_write_softc;
  1047. hw_raw_write_mapping[0xe] = csky_pmu_write_cbmc;
  1048. hw_raw_write_mapping[0xf] = csky_pmu_write_cbic;
  1049. hw_raw_write_mapping[0x10] = csky_pmu_write_ibmc;
  1050. hw_raw_write_mapping[0x11] = csky_pmu_write_ibic;
  1051. hw_raw_write_mapping[0x12] = csky_pmu_write_lsfc;
  1052. hw_raw_write_mapping[0x13] = csky_pmu_write_sic;
  1053. hw_raw_write_mapping[0x14] = csky_pmu_write_dcrac;
  1054. hw_raw_write_mapping[0x15] = csky_pmu_write_dcrmc;
  1055. hw_raw_write_mapping[0x16] = csky_pmu_write_dcwac;
  1056. hw_raw_write_mapping[0x17] = csky_pmu_write_dcwmc;
  1057. hw_raw_write_mapping[0x18] = csky_pmu_write_l2rac;
  1058. hw_raw_write_mapping[0x19] = csky_pmu_write_l2rmc;
  1059. hw_raw_write_mapping[0x1a] = csky_pmu_write_l2wac;
  1060. hw_raw_write_mapping[0x1b] = csky_pmu_write_l2wmc;
  1061. return 0;
  1062. }
  1063. static int csky_pmu_starting_cpu(unsigned int cpu)
  1064. {
  1065. enable_percpu_irq(csky_pmu_irq, 0);
  1066. return 0;
  1067. }
  1068. static int csky_pmu_dying_cpu(unsigned int cpu)
  1069. {
  1070. disable_percpu_irq(csky_pmu_irq);
  1071. return 0;
  1072. }
  1073. int csky_pmu_device_probe(struct platform_device *pdev,
  1074. const struct of_device_id *of_table)
  1075. {
  1076. struct device_node *node = pdev->dev.of_node;
  1077. int ret;
  1078. ret = init_hw_perf_events();
  1079. if (ret) {
  1080. pr_notice("[perf] failed to probe PMU!\n");
  1081. return ret;
  1082. }
  1083. if (of_property_read_u32(node, "count-width",
  1084. &csky_pmu.count_width)) {
  1085. csky_pmu.count_width = DEFAULT_COUNT_WIDTH;
  1086. }
  1087. csky_pmu.max_period = BIT_ULL(csky_pmu.count_width) - 1;
  1088. csky_pmu.plat_device = pdev;
  1089. /* Ensure the PMU has sane values out of reset. */
  1090. on_each_cpu(csky_pmu_reset, &csky_pmu, 1);
  1091. ret = csky_pmu_request_irq(csky_pmu_handle_irq);
  1092. if (ret) {
  1093. csky_pmu.pmu.capabilities |= PERF_PMU_CAP_NO_INTERRUPT;
  1094. pr_notice("[perf] PMU request irq fail!\n");
  1095. }
  1096. ret = cpuhp_setup_state(CPUHP_AP_PERF_CSKY_ONLINE, "AP_PERF_ONLINE",
  1097. csky_pmu_starting_cpu,
  1098. csky_pmu_dying_cpu);
  1099. if (ret) {
  1100. csky_pmu_free_irq();
  1101. free_percpu(csky_pmu.hw_events);
  1102. return ret;
  1103. }
  1104. ret = perf_pmu_register(&csky_pmu.pmu, "cpu", PERF_TYPE_RAW);
  1105. if (ret) {
  1106. csky_pmu_free_irq();
  1107. free_percpu(csky_pmu.hw_events);
  1108. }
  1109. return ret;
  1110. }
  1111. static const struct of_device_id csky_pmu_of_device_ids[] = {
  1112. {.compatible = "csky,csky-pmu"},
  1113. {},
  1114. };
  1115. static int csky_pmu_dev_probe(struct platform_device *pdev)
  1116. {
  1117. return csky_pmu_device_probe(pdev, csky_pmu_of_device_ids);
  1118. }
  1119. static struct platform_driver csky_pmu_driver = {
  1120. .driver = {
  1121. .name = "csky-pmu",
  1122. .of_match_table = csky_pmu_of_device_ids,
  1123. },
  1124. .probe = csky_pmu_dev_probe,
  1125. };
  1126. static int __init csky_pmu_probe(void)
  1127. {
  1128. int ret;
  1129. ret = platform_driver_register(&csky_pmu_driver);
  1130. if (ret)
  1131. pr_notice("[perf] PMU initialization failed\n");
  1132. else
  1133. pr_notice("[perf] PMU initialization done\n");
  1134. return ret;
  1135. }
  1136. device_initcall(csky_pmu_probe);