marvell_cn10k_ddr_pmu.c 22 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758
  1. // SPDX-License-Identifier: GPL-2.0
  2. /* Marvell CN10K DRAM Subsystem (DSS) Performance Monitor Driver
  3. *
  4. * Copyright (C) 2021 Marvell.
  5. */
  6. #include <linux/init.h>
  7. #include <linux/io.h>
  8. #include <linux/module.h>
  9. #include <linux/of.h>
  10. #include <linux/of_address.h>
  11. #include <linux/of_device.h>
  12. #include <linux/perf_event.h>
  13. #include <linux/hrtimer.h>
  14. /* Performance Counters Operating Mode Control Registers */
  15. #define DDRC_PERF_CNT_OP_MODE_CTRL 0x8020
  16. #define OP_MODE_CTRL_VAL_MANNUAL 0x1
  17. /* Performance Counters Start Operation Control Registers */
  18. #define DDRC_PERF_CNT_START_OP_CTRL 0x8028
  19. #define START_OP_CTRL_VAL_START 0x1ULL
  20. #define START_OP_CTRL_VAL_ACTIVE 0x2
  21. /* Performance Counters End Operation Control Registers */
  22. #define DDRC_PERF_CNT_END_OP_CTRL 0x8030
  23. #define END_OP_CTRL_VAL_END 0x1ULL
  24. /* Performance Counters End Status Registers */
  25. #define DDRC_PERF_CNT_END_STATUS 0x8038
  26. #define END_STATUS_VAL_END_TIMER_MODE_END 0x1
  27. /* Performance Counters Configuration Registers */
  28. #define DDRC_PERF_CFG_BASE 0x8040
  29. /* 8 Generic event counter + 2 fixed event counters */
  30. #define DDRC_PERF_NUM_GEN_COUNTERS 8
  31. #define DDRC_PERF_NUM_FIX_COUNTERS 2
  32. #define DDRC_PERF_READ_COUNTER_IDX DDRC_PERF_NUM_GEN_COUNTERS
  33. #define DDRC_PERF_WRITE_COUNTER_IDX (DDRC_PERF_NUM_GEN_COUNTERS + 1)
  34. #define DDRC_PERF_NUM_COUNTERS (DDRC_PERF_NUM_GEN_COUNTERS + \
  35. DDRC_PERF_NUM_FIX_COUNTERS)
  36. /* Generic event counter registers */
  37. #define DDRC_PERF_CFG(n) (DDRC_PERF_CFG_BASE + 8 * (n))
  38. #define EVENT_ENABLE BIT_ULL(63)
  39. /* Two dedicated event counters for DDR reads and writes */
  40. #define EVENT_DDR_READS 101
  41. #define EVENT_DDR_WRITES 100
  42. /*
  43. * programmable events IDs in programmable event counters.
  44. * DO NOT change these event-id numbers, they are used to
  45. * program event bitmap in h/w.
  46. */
  47. #define EVENT_OP_IS_ZQLATCH 55
  48. #define EVENT_OP_IS_ZQSTART 54
  49. #define EVENT_OP_IS_TCR_MRR 53
  50. #define EVENT_OP_IS_DQSOSC_MRR 52
  51. #define EVENT_OP_IS_DQSOSC_MPC 51
  52. #define EVENT_VISIBLE_WIN_LIMIT_REACHED_WR 50
  53. #define EVENT_VISIBLE_WIN_LIMIT_REACHED_RD 49
  54. #define EVENT_BSM_STARVATION 48
  55. #define EVENT_BSM_ALLOC 47
  56. #define EVENT_LPR_REQ_WITH_NOCREDIT 46
  57. #define EVENT_HPR_REQ_WITH_NOCREDIT 45
  58. #define EVENT_OP_IS_ZQCS 44
  59. #define EVENT_OP_IS_ZQCL 43
  60. #define EVENT_OP_IS_LOAD_MODE 42
  61. #define EVENT_OP_IS_SPEC_REF 41
  62. #define EVENT_OP_IS_CRIT_REF 40
  63. #define EVENT_OP_IS_REFRESH 39
  64. #define EVENT_OP_IS_ENTER_MPSM 35
  65. #define EVENT_OP_IS_ENTER_POWERDOWN 31
  66. #define EVENT_OP_IS_ENTER_SELFREF 27
  67. #define EVENT_WAW_HAZARD 26
  68. #define EVENT_RAW_HAZARD 25
  69. #define EVENT_WAR_HAZARD 24
  70. #define EVENT_WRITE_COMBINE 23
  71. #define EVENT_RDWR_TRANSITIONS 22
  72. #define EVENT_PRECHARGE_FOR_OTHER 21
  73. #define EVENT_PRECHARGE_FOR_RDWR 20
  74. #define EVENT_OP_IS_PRECHARGE 19
  75. #define EVENT_OP_IS_MWR 18
  76. #define EVENT_OP_IS_WR 17
  77. #define EVENT_OP_IS_RD 16
  78. #define EVENT_OP_IS_RD_ACTIVATE 15
  79. #define EVENT_OP_IS_RD_OR_WR 14
  80. #define EVENT_OP_IS_ACTIVATE 13
  81. #define EVENT_WR_XACT_WHEN_CRITICAL 12
  82. #define EVENT_LPR_XACT_WHEN_CRITICAL 11
  83. #define EVENT_HPR_XACT_WHEN_CRITICAL 10
  84. #define EVENT_DFI_RD_DATA_CYCLES 9
  85. #define EVENT_DFI_WR_DATA_CYCLES 8
  86. #define EVENT_ACT_BYPASS 7
  87. #define EVENT_READ_BYPASS 6
  88. #define EVENT_HIF_HI_PRI_RD 5
  89. #define EVENT_HIF_RMW 4
  90. #define EVENT_HIF_RD 3
  91. #define EVENT_HIF_WR 2
  92. #define EVENT_HIF_RD_OR_WR 1
  93. /* Event counter value registers */
  94. #define DDRC_PERF_CNT_VALUE_BASE 0x8080
  95. #define DDRC_PERF_CNT_VALUE(n) (DDRC_PERF_CNT_VALUE_BASE + 8 * (n))
  96. /* Fixed event counter enable/disable register */
  97. #define DDRC_PERF_CNT_FREERUN_EN 0x80C0
  98. #define DDRC_PERF_FREERUN_WRITE_EN 0x1
  99. #define DDRC_PERF_FREERUN_READ_EN 0x2
  100. /* Fixed event counter control register */
  101. #define DDRC_PERF_CNT_FREERUN_CTRL 0x80C8
  102. #define DDRC_FREERUN_WRITE_CNT_CLR 0x1
  103. #define DDRC_FREERUN_READ_CNT_CLR 0x2
  104. /* Fixed event counter value register */
  105. #define DDRC_PERF_CNT_VALUE_WR_OP 0x80D0
  106. #define DDRC_PERF_CNT_VALUE_RD_OP 0x80D8
  107. #define DDRC_PERF_CNT_VALUE_OVERFLOW BIT_ULL(48)
  108. #define DDRC_PERF_CNT_MAX_VALUE GENMASK_ULL(48, 0)
  109. struct cn10k_ddr_pmu {
  110. struct pmu pmu;
  111. void __iomem *base;
  112. unsigned int cpu;
  113. struct device *dev;
  114. int active_events;
  115. struct perf_event *events[DDRC_PERF_NUM_COUNTERS];
  116. struct hrtimer hrtimer;
  117. struct hlist_node node;
  118. };
  119. #define to_cn10k_ddr_pmu(p) container_of(p, struct cn10k_ddr_pmu, pmu)
  120. static ssize_t cn10k_ddr_pmu_event_show(struct device *dev,
  121. struct device_attribute *attr,
  122. char *page)
  123. {
  124. struct perf_pmu_events_attr *pmu_attr;
  125. pmu_attr = container_of(attr, struct perf_pmu_events_attr, attr);
  126. return sysfs_emit(page, "event=0x%02llx\n", pmu_attr->id);
  127. }
  128. #define CN10K_DDR_PMU_EVENT_ATTR(_name, _id) \
  129. PMU_EVENT_ATTR_ID(_name, cn10k_ddr_pmu_event_show, _id)
  130. static struct attribute *cn10k_ddr_perf_events_attrs[] = {
  131. CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_rd_or_wr_access, EVENT_HIF_RD_OR_WR),
  132. CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_wr_access, EVENT_HIF_WR),
  133. CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_rd_access, EVENT_HIF_RD),
  134. CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_rmw_access, EVENT_HIF_RMW),
  135. CN10K_DDR_PMU_EVENT_ATTR(ddr_hif_pri_rdaccess, EVENT_HIF_HI_PRI_RD),
  136. CN10K_DDR_PMU_EVENT_ATTR(ddr_rd_bypass_access, EVENT_READ_BYPASS),
  137. CN10K_DDR_PMU_EVENT_ATTR(ddr_act_bypass_access, EVENT_ACT_BYPASS),
  138. CN10K_DDR_PMU_EVENT_ATTR(ddr_dif_wr_data_access, EVENT_DFI_WR_DATA_CYCLES),
  139. CN10K_DDR_PMU_EVENT_ATTR(ddr_dif_rd_data_access, EVENT_DFI_RD_DATA_CYCLES),
  140. CN10K_DDR_PMU_EVENT_ATTR(ddr_hpri_sched_rd_crit_access,
  141. EVENT_HPR_XACT_WHEN_CRITICAL),
  142. CN10K_DDR_PMU_EVENT_ATTR(ddr_lpri_sched_rd_crit_access,
  143. EVENT_LPR_XACT_WHEN_CRITICAL),
  144. CN10K_DDR_PMU_EVENT_ATTR(ddr_wr_trxn_crit_access,
  145. EVENT_WR_XACT_WHEN_CRITICAL),
  146. CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_active_access, EVENT_OP_IS_ACTIVATE),
  147. CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_rd_or_wr_access, EVENT_OP_IS_RD_OR_WR),
  148. CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_rd_active_access, EVENT_OP_IS_RD_ACTIVATE),
  149. CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_read, EVENT_OP_IS_RD),
  150. CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_write, EVENT_OP_IS_WR),
  151. CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_mwr, EVENT_OP_IS_MWR),
  152. CN10K_DDR_PMU_EVENT_ATTR(ddr_precharge, EVENT_OP_IS_PRECHARGE),
  153. CN10K_DDR_PMU_EVENT_ATTR(ddr_precharge_for_rdwr, EVENT_PRECHARGE_FOR_RDWR),
  154. CN10K_DDR_PMU_EVENT_ATTR(ddr_precharge_for_other,
  155. EVENT_PRECHARGE_FOR_OTHER),
  156. CN10K_DDR_PMU_EVENT_ATTR(ddr_rdwr_transitions, EVENT_RDWR_TRANSITIONS),
  157. CN10K_DDR_PMU_EVENT_ATTR(ddr_write_combine, EVENT_WRITE_COMBINE),
  158. CN10K_DDR_PMU_EVENT_ATTR(ddr_war_hazard, EVENT_WAR_HAZARD),
  159. CN10K_DDR_PMU_EVENT_ATTR(ddr_raw_hazard, EVENT_RAW_HAZARD),
  160. CN10K_DDR_PMU_EVENT_ATTR(ddr_waw_hazard, EVENT_WAW_HAZARD),
  161. CN10K_DDR_PMU_EVENT_ATTR(ddr_enter_selfref, EVENT_OP_IS_ENTER_SELFREF),
  162. CN10K_DDR_PMU_EVENT_ATTR(ddr_enter_powerdown, EVENT_OP_IS_ENTER_POWERDOWN),
  163. CN10K_DDR_PMU_EVENT_ATTR(ddr_enter_mpsm, EVENT_OP_IS_ENTER_MPSM),
  164. CN10K_DDR_PMU_EVENT_ATTR(ddr_refresh, EVENT_OP_IS_REFRESH),
  165. CN10K_DDR_PMU_EVENT_ATTR(ddr_crit_ref, EVENT_OP_IS_CRIT_REF),
  166. CN10K_DDR_PMU_EVENT_ATTR(ddr_spec_ref, EVENT_OP_IS_SPEC_REF),
  167. CN10K_DDR_PMU_EVENT_ATTR(ddr_load_mode, EVENT_OP_IS_LOAD_MODE),
  168. CN10K_DDR_PMU_EVENT_ATTR(ddr_zqcl, EVENT_OP_IS_ZQCL),
  169. CN10K_DDR_PMU_EVENT_ATTR(ddr_cam_wr_access, EVENT_OP_IS_ZQCS),
  170. CN10K_DDR_PMU_EVENT_ATTR(ddr_hpr_req_with_nocredit,
  171. EVENT_HPR_REQ_WITH_NOCREDIT),
  172. CN10K_DDR_PMU_EVENT_ATTR(ddr_lpr_req_with_nocredit,
  173. EVENT_LPR_REQ_WITH_NOCREDIT),
  174. CN10K_DDR_PMU_EVENT_ATTR(ddr_bsm_alloc, EVENT_BSM_ALLOC),
  175. CN10K_DDR_PMU_EVENT_ATTR(ddr_bsm_starvation, EVENT_BSM_STARVATION),
  176. CN10K_DDR_PMU_EVENT_ATTR(ddr_win_limit_reached_rd,
  177. EVENT_VISIBLE_WIN_LIMIT_REACHED_RD),
  178. CN10K_DDR_PMU_EVENT_ATTR(ddr_win_limit_reached_wr,
  179. EVENT_VISIBLE_WIN_LIMIT_REACHED_WR),
  180. CN10K_DDR_PMU_EVENT_ATTR(ddr_dqsosc_mpc, EVENT_OP_IS_DQSOSC_MPC),
  181. CN10K_DDR_PMU_EVENT_ATTR(ddr_dqsosc_mrr, EVENT_OP_IS_DQSOSC_MRR),
  182. CN10K_DDR_PMU_EVENT_ATTR(ddr_tcr_mrr, EVENT_OP_IS_TCR_MRR),
  183. CN10K_DDR_PMU_EVENT_ATTR(ddr_zqstart, EVENT_OP_IS_ZQSTART),
  184. CN10K_DDR_PMU_EVENT_ATTR(ddr_zqlatch, EVENT_OP_IS_ZQLATCH),
  185. /* Free run event counters */
  186. CN10K_DDR_PMU_EVENT_ATTR(ddr_ddr_reads, EVENT_DDR_READS),
  187. CN10K_DDR_PMU_EVENT_ATTR(ddr_ddr_writes, EVENT_DDR_WRITES),
  188. NULL
  189. };
  190. static struct attribute_group cn10k_ddr_perf_events_attr_group = {
  191. .name = "events",
  192. .attrs = cn10k_ddr_perf_events_attrs,
  193. };
  194. PMU_FORMAT_ATTR(event, "config:0-8");
  195. static struct attribute *cn10k_ddr_perf_format_attrs[] = {
  196. &format_attr_event.attr,
  197. NULL,
  198. };
  199. static struct attribute_group cn10k_ddr_perf_format_attr_group = {
  200. .name = "format",
  201. .attrs = cn10k_ddr_perf_format_attrs,
  202. };
  203. static ssize_t cn10k_ddr_perf_cpumask_show(struct device *dev,
  204. struct device_attribute *attr,
  205. char *buf)
  206. {
  207. struct cn10k_ddr_pmu *pmu = dev_get_drvdata(dev);
  208. return cpumap_print_to_pagebuf(true, buf, cpumask_of(pmu->cpu));
  209. }
  210. static struct device_attribute cn10k_ddr_perf_cpumask_attr =
  211. __ATTR(cpumask, 0444, cn10k_ddr_perf_cpumask_show, NULL);
  212. static struct attribute *cn10k_ddr_perf_cpumask_attrs[] = {
  213. &cn10k_ddr_perf_cpumask_attr.attr,
  214. NULL,
  215. };
  216. static struct attribute_group cn10k_ddr_perf_cpumask_attr_group = {
  217. .attrs = cn10k_ddr_perf_cpumask_attrs,
  218. };
  219. static const struct attribute_group *cn10k_attr_groups[] = {
  220. &cn10k_ddr_perf_events_attr_group,
  221. &cn10k_ddr_perf_format_attr_group,
  222. &cn10k_ddr_perf_cpumask_attr_group,
  223. NULL,
  224. };
  225. /* Default poll timeout is 100 sec, which is very sufficient for
  226. * 48 bit counter incremented max at 5.6 GT/s, which may take many
  227. * hours to overflow.
  228. */
  229. static unsigned long cn10k_ddr_pmu_poll_period_sec = 100;
  230. module_param_named(poll_period_sec, cn10k_ddr_pmu_poll_period_sec, ulong, 0644);
  231. static ktime_t cn10k_ddr_pmu_timer_period(void)
  232. {
  233. return ms_to_ktime((u64)cn10k_ddr_pmu_poll_period_sec * USEC_PER_SEC);
  234. }
  235. static int ddr_perf_get_event_bitmap(int eventid, u64 *event_bitmap)
  236. {
  237. switch (eventid) {
  238. case EVENT_HIF_RD_OR_WR ... EVENT_WAW_HAZARD:
  239. case EVENT_OP_IS_REFRESH ... EVENT_OP_IS_ZQLATCH:
  240. *event_bitmap = (1ULL << (eventid - 1));
  241. break;
  242. case EVENT_OP_IS_ENTER_SELFREF:
  243. case EVENT_OP_IS_ENTER_POWERDOWN:
  244. case EVENT_OP_IS_ENTER_MPSM:
  245. *event_bitmap = (0xFULL << (eventid - 1));
  246. break;
  247. default:
  248. pr_err("%s Invalid eventid %d\n", __func__, eventid);
  249. return -EINVAL;
  250. }
  251. return 0;
  252. }
  253. static int cn10k_ddr_perf_alloc_counter(struct cn10k_ddr_pmu *pmu,
  254. struct perf_event *event)
  255. {
  256. u8 config = event->attr.config;
  257. int i;
  258. /* DDR read free-run counter index */
  259. if (config == EVENT_DDR_READS) {
  260. pmu->events[DDRC_PERF_READ_COUNTER_IDX] = event;
  261. return DDRC_PERF_READ_COUNTER_IDX;
  262. }
  263. /* DDR write free-run counter index */
  264. if (config == EVENT_DDR_WRITES) {
  265. pmu->events[DDRC_PERF_WRITE_COUNTER_IDX] = event;
  266. return DDRC_PERF_WRITE_COUNTER_IDX;
  267. }
  268. /* Allocate DDR generic counters */
  269. for (i = 0; i < DDRC_PERF_NUM_GEN_COUNTERS; i++) {
  270. if (pmu->events[i] == NULL) {
  271. pmu->events[i] = event;
  272. return i;
  273. }
  274. }
  275. return -ENOENT;
  276. }
  277. static void cn10k_ddr_perf_free_counter(struct cn10k_ddr_pmu *pmu, int counter)
  278. {
  279. pmu->events[counter] = NULL;
  280. }
  281. static int cn10k_ddr_perf_event_init(struct perf_event *event)
  282. {
  283. struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu);
  284. struct hw_perf_event *hwc = &event->hw;
  285. if (event->attr.type != event->pmu->type)
  286. return -ENOENT;
  287. if (is_sampling_event(event)) {
  288. dev_info(pmu->dev, "Sampling not supported!\n");
  289. return -EOPNOTSUPP;
  290. }
  291. if (event->cpu < 0) {
  292. dev_warn(pmu->dev, "Can't provide per-task data!\n");
  293. return -EOPNOTSUPP;
  294. }
  295. /* We must NOT create groups containing mixed PMUs */
  296. if (event->group_leader->pmu != event->pmu &&
  297. !is_software_event(event->group_leader))
  298. return -EINVAL;
  299. /* Set ownership of event to one CPU, same event can not be observed
  300. * on multiple cpus at same time.
  301. */
  302. event->cpu = pmu->cpu;
  303. hwc->idx = -1;
  304. return 0;
  305. }
  306. static void cn10k_ddr_perf_counter_enable(struct cn10k_ddr_pmu *pmu,
  307. int counter, bool enable)
  308. {
  309. u32 reg;
  310. u64 val;
  311. if (counter > DDRC_PERF_NUM_COUNTERS) {
  312. pr_err("Error: unsupported counter %d\n", counter);
  313. return;
  314. }
  315. if (counter < DDRC_PERF_NUM_GEN_COUNTERS) {
  316. reg = DDRC_PERF_CFG(counter);
  317. val = readq_relaxed(pmu->base + reg);
  318. if (enable)
  319. val |= EVENT_ENABLE;
  320. else
  321. val &= ~EVENT_ENABLE;
  322. writeq_relaxed(val, pmu->base + reg);
  323. } else {
  324. val = readq_relaxed(pmu->base + DDRC_PERF_CNT_FREERUN_EN);
  325. if (enable) {
  326. if (counter == DDRC_PERF_READ_COUNTER_IDX)
  327. val |= DDRC_PERF_FREERUN_READ_EN;
  328. else
  329. val |= DDRC_PERF_FREERUN_WRITE_EN;
  330. } else {
  331. if (counter == DDRC_PERF_READ_COUNTER_IDX)
  332. val &= ~DDRC_PERF_FREERUN_READ_EN;
  333. else
  334. val &= ~DDRC_PERF_FREERUN_WRITE_EN;
  335. }
  336. writeq_relaxed(val, pmu->base + DDRC_PERF_CNT_FREERUN_EN);
  337. }
  338. }
  339. static u64 cn10k_ddr_perf_read_counter(struct cn10k_ddr_pmu *pmu, int counter)
  340. {
  341. u64 val;
  342. if (counter == DDRC_PERF_READ_COUNTER_IDX)
  343. return readq_relaxed(pmu->base + DDRC_PERF_CNT_VALUE_RD_OP);
  344. if (counter == DDRC_PERF_WRITE_COUNTER_IDX)
  345. return readq_relaxed(pmu->base + DDRC_PERF_CNT_VALUE_WR_OP);
  346. val = readq_relaxed(pmu->base + DDRC_PERF_CNT_VALUE(counter));
  347. return val;
  348. }
  349. static void cn10k_ddr_perf_event_update(struct perf_event *event)
  350. {
  351. struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu);
  352. struct hw_perf_event *hwc = &event->hw;
  353. u64 prev_count, new_count, mask;
  354. do {
  355. prev_count = local64_read(&hwc->prev_count);
  356. new_count = cn10k_ddr_perf_read_counter(pmu, hwc->idx);
  357. } while (local64_xchg(&hwc->prev_count, new_count) != prev_count);
  358. mask = DDRC_PERF_CNT_MAX_VALUE;
  359. local64_add((new_count - prev_count) & mask, &event->count);
  360. }
  361. static void cn10k_ddr_perf_event_start(struct perf_event *event, int flags)
  362. {
  363. struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu);
  364. struct hw_perf_event *hwc = &event->hw;
  365. int counter = hwc->idx;
  366. local64_set(&hwc->prev_count, 0);
  367. cn10k_ddr_perf_counter_enable(pmu, counter, true);
  368. hwc->state = 0;
  369. }
  370. static int cn10k_ddr_perf_event_add(struct perf_event *event, int flags)
  371. {
  372. struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu);
  373. struct hw_perf_event *hwc = &event->hw;
  374. u8 config = event->attr.config;
  375. int counter, ret;
  376. u32 reg_offset;
  377. u64 val;
  378. counter = cn10k_ddr_perf_alloc_counter(pmu, event);
  379. if (counter < 0)
  380. return -EAGAIN;
  381. pmu->active_events++;
  382. hwc->idx = counter;
  383. if (pmu->active_events == 1)
  384. hrtimer_start(&pmu->hrtimer, cn10k_ddr_pmu_timer_period(),
  385. HRTIMER_MODE_REL_PINNED);
  386. if (counter < DDRC_PERF_NUM_GEN_COUNTERS) {
  387. /* Generic counters, configure event id */
  388. reg_offset = DDRC_PERF_CFG(counter);
  389. ret = ddr_perf_get_event_bitmap(config, &val);
  390. if (ret)
  391. return ret;
  392. writeq_relaxed(val, pmu->base + reg_offset);
  393. } else {
  394. /* fixed event counter, clear counter value */
  395. if (counter == DDRC_PERF_READ_COUNTER_IDX)
  396. val = DDRC_FREERUN_READ_CNT_CLR;
  397. else
  398. val = DDRC_FREERUN_WRITE_CNT_CLR;
  399. writeq_relaxed(val, pmu->base + DDRC_PERF_CNT_FREERUN_CTRL);
  400. }
  401. hwc->state |= PERF_HES_STOPPED;
  402. if (flags & PERF_EF_START)
  403. cn10k_ddr_perf_event_start(event, flags);
  404. return 0;
  405. }
  406. static void cn10k_ddr_perf_event_stop(struct perf_event *event, int flags)
  407. {
  408. struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu);
  409. struct hw_perf_event *hwc = &event->hw;
  410. int counter = hwc->idx;
  411. cn10k_ddr_perf_counter_enable(pmu, counter, false);
  412. if (flags & PERF_EF_UPDATE)
  413. cn10k_ddr_perf_event_update(event);
  414. hwc->state |= PERF_HES_STOPPED;
  415. }
  416. static void cn10k_ddr_perf_event_del(struct perf_event *event, int flags)
  417. {
  418. struct cn10k_ddr_pmu *pmu = to_cn10k_ddr_pmu(event->pmu);
  419. struct hw_perf_event *hwc = &event->hw;
  420. int counter = hwc->idx;
  421. cn10k_ddr_perf_event_stop(event, PERF_EF_UPDATE);
  422. cn10k_ddr_perf_free_counter(pmu, counter);
  423. pmu->active_events--;
  424. hwc->idx = -1;
  425. /* Cancel timer when no events to capture */
  426. if (pmu->active_events == 0)
  427. hrtimer_cancel(&pmu->hrtimer);
  428. }
  429. static void cn10k_ddr_perf_pmu_enable(struct pmu *pmu)
  430. {
  431. struct cn10k_ddr_pmu *ddr_pmu = to_cn10k_ddr_pmu(pmu);
  432. writeq_relaxed(START_OP_CTRL_VAL_START, ddr_pmu->base +
  433. DDRC_PERF_CNT_START_OP_CTRL);
  434. }
  435. static void cn10k_ddr_perf_pmu_disable(struct pmu *pmu)
  436. {
  437. struct cn10k_ddr_pmu *ddr_pmu = to_cn10k_ddr_pmu(pmu);
  438. writeq_relaxed(END_OP_CTRL_VAL_END, ddr_pmu->base +
  439. DDRC_PERF_CNT_END_OP_CTRL);
  440. }
  441. static void cn10k_ddr_perf_event_update_all(struct cn10k_ddr_pmu *pmu)
  442. {
  443. struct hw_perf_event *hwc;
  444. int i;
  445. for (i = 0; i < DDRC_PERF_NUM_GEN_COUNTERS; i++) {
  446. if (pmu->events[i] == NULL)
  447. continue;
  448. cn10k_ddr_perf_event_update(pmu->events[i]);
  449. }
  450. /* Reset previous count as h/w counter are reset */
  451. for (i = 0; i < DDRC_PERF_NUM_GEN_COUNTERS; i++) {
  452. if (pmu->events[i] == NULL)
  453. continue;
  454. hwc = &pmu->events[i]->hw;
  455. local64_set(&hwc->prev_count, 0);
  456. }
  457. }
  458. static irqreturn_t cn10k_ddr_pmu_overflow_handler(struct cn10k_ddr_pmu *pmu)
  459. {
  460. struct perf_event *event;
  461. struct hw_perf_event *hwc;
  462. u64 prev_count, new_count;
  463. u64 value;
  464. int i;
  465. event = pmu->events[DDRC_PERF_READ_COUNTER_IDX];
  466. if (event) {
  467. hwc = &event->hw;
  468. prev_count = local64_read(&hwc->prev_count);
  469. new_count = cn10k_ddr_perf_read_counter(pmu, hwc->idx);
  470. /* Overflow condition is when new count less than
  471. * previous count
  472. */
  473. if (new_count < prev_count)
  474. cn10k_ddr_perf_event_update(event);
  475. }
  476. event = pmu->events[DDRC_PERF_WRITE_COUNTER_IDX];
  477. if (event) {
  478. hwc = &event->hw;
  479. prev_count = local64_read(&hwc->prev_count);
  480. new_count = cn10k_ddr_perf_read_counter(pmu, hwc->idx);
  481. /* Overflow condition is when new count less than
  482. * previous count
  483. */
  484. if (new_count < prev_count)
  485. cn10k_ddr_perf_event_update(event);
  486. }
  487. for (i = 0; i < DDRC_PERF_NUM_GEN_COUNTERS; i++) {
  488. if (pmu->events[i] == NULL)
  489. continue;
  490. value = cn10k_ddr_perf_read_counter(pmu, i);
  491. if (value == DDRC_PERF_CNT_MAX_VALUE) {
  492. pr_info("Counter-(%d) reached max value\n", i);
  493. cn10k_ddr_perf_event_update_all(pmu);
  494. cn10k_ddr_perf_pmu_disable(&pmu->pmu);
  495. cn10k_ddr_perf_pmu_enable(&pmu->pmu);
  496. }
  497. }
  498. return IRQ_HANDLED;
  499. }
  500. static enum hrtimer_restart cn10k_ddr_pmu_timer_handler(struct hrtimer *hrtimer)
  501. {
  502. struct cn10k_ddr_pmu *pmu = container_of(hrtimer, struct cn10k_ddr_pmu,
  503. hrtimer);
  504. unsigned long flags;
  505. local_irq_save(flags);
  506. cn10k_ddr_pmu_overflow_handler(pmu);
  507. local_irq_restore(flags);
  508. hrtimer_forward_now(hrtimer, cn10k_ddr_pmu_timer_period());
  509. return HRTIMER_RESTART;
  510. }
  511. static int cn10k_ddr_pmu_offline_cpu(unsigned int cpu, struct hlist_node *node)
  512. {
  513. struct cn10k_ddr_pmu *pmu = hlist_entry_safe(node, struct cn10k_ddr_pmu,
  514. node);
  515. unsigned int target;
  516. if (cpu != pmu->cpu)
  517. return 0;
  518. target = cpumask_any_but(cpu_online_mask, cpu);
  519. if (target >= nr_cpu_ids)
  520. return 0;
  521. perf_pmu_migrate_context(&pmu->pmu, cpu, target);
  522. pmu->cpu = target;
  523. return 0;
  524. }
  525. static int cn10k_ddr_perf_probe(struct platform_device *pdev)
  526. {
  527. struct cn10k_ddr_pmu *ddr_pmu;
  528. struct resource *res;
  529. void __iomem *base;
  530. char *name;
  531. int ret;
  532. ddr_pmu = devm_kzalloc(&pdev->dev, sizeof(*ddr_pmu), GFP_KERNEL);
  533. if (!ddr_pmu)
  534. return -ENOMEM;
  535. ddr_pmu->dev = &pdev->dev;
  536. platform_set_drvdata(pdev, ddr_pmu);
  537. base = devm_platform_get_and_ioremap_resource(pdev, 0, &res);
  538. if (IS_ERR(base))
  539. return PTR_ERR(base);
  540. ddr_pmu->base = base;
  541. /* Setup the PMU counter to work in manual mode */
  542. writeq_relaxed(OP_MODE_CTRL_VAL_MANNUAL, ddr_pmu->base +
  543. DDRC_PERF_CNT_OP_MODE_CTRL);
  544. ddr_pmu->pmu = (struct pmu) {
  545. .module = THIS_MODULE,
  546. .capabilities = PERF_PMU_CAP_NO_EXCLUDE,
  547. .task_ctx_nr = perf_invalid_context,
  548. .attr_groups = cn10k_attr_groups,
  549. .event_init = cn10k_ddr_perf_event_init,
  550. .add = cn10k_ddr_perf_event_add,
  551. .del = cn10k_ddr_perf_event_del,
  552. .start = cn10k_ddr_perf_event_start,
  553. .stop = cn10k_ddr_perf_event_stop,
  554. .read = cn10k_ddr_perf_event_update,
  555. .pmu_enable = cn10k_ddr_perf_pmu_enable,
  556. .pmu_disable = cn10k_ddr_perf_pmu_disable,
  557. };
  558. /* Choose this cpu to collect perf data */
  559. ddr_pmu->cpu = raw_smp_processor_id();
  560. name = devm_kasprintf(ddr_pmu->dev, GFP_KERNEL, "mrvl_ddr_pmu_%llx",
  561. res->start);
  562. if (!name)
  563. return -ENOMEM;
  564. hrtimer_init(&ddr_pmu->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
  565. ddr_pmu->hrtimer.function = cn10k_ddr_pmu_timer_handler;
  566. cpuhp_state_add_instance_nocalls(
  567. CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE,
  568. &ddr_pmu->node);
  569. ret = perf_pmu_register(&ddr_pmu->pmu, name, -1);
  570. if (ret)
  571. goto error;
  572. pr_info("CN10K DDR PMU Driver for ddrc@%llx\n", res->start);
  573. return 0;
  574. error:
  575. cpuhp_state_remove_instance_nocalls(
  576. CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE,
  577. &ddr_pmu->node);
  578. return ret;
  579. }
  580. static int cn10k_ddr_perf_remove(struct platform_device *pdev)
  581. {
  582. struct cn10k_ddr_pmu *ddr_pmu = platform_get_drvdata(pdev);
  583. cpuhp_state_remove_instance_nocalls(
  584. CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE,
  585. &ddr_pmu->node);
  586. perf_pmu_unregister(&ddr_pmu->pmu);
  587. return 0;
  588. }
  589. #ifdef CONFIG_OF
  590. static const struct of_device_id cn10k_ddr_pmu_of_match[] = {
  591. { .compatible = "marvell,cn10k-ddr-pmu", },
  592. { },
  593. };
  594. MODULE_DEVICE_TABLE(of, cn10k_ddr_pmu_of_match);
  595. #endif
  596. static struct platform_driver cn10k_ddr_pmu_driver = {
  597. .driver = {
  598. .name = "cn10k-ddr-pmu",
  599. .of_match_table = of_match_ptr(cn10k_ddr_pmu_of_match),
  600. .suppress_bind_attrs = true,
  601. },
  602. .probe = cn10k_ddr_perf_probe,
  603. .remove = cn10k_ddr_perf_remove,
  604. };
  605. static int __init cn10k_ddr_pmu_init(void)
  606. {
  607. int ret;
  608. ret = cpuhp_setup_state_multi(
  609. CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE,
  610. "perf/marvell/cn10k/ddr:online", NULL,
  611. cn10k_ddr_pmu_offline_cpu);
  612. if (ret)
  613. return ret;
  614. ret = platform_driver_register(&cn10k_ddr_pmu_driver);
  615. if (ret)
  616. cpuhp_remove_multi_state(
  617. CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE);
  618. return ret;
  619. }
  620. static void __exit cn10k_ddr_pmu_exit(void)
  621. {
  622. platform_driver_unregister(&cn10k_ddr_pmu_driver);
  623. cpuhp_remove_multi_state(CPUHP_AP_PERF_ARM_MARVELL_CN10K_DDR_ONLINE);
  624. }
  625. module_init(cn10k_ddr_pmu_init);
  626. module_exit(cn10k_ddr_pmu_exit);
  627. MODULE_AUTHOR("Bharat Bhushan <[email protected]>");
  628. MODULE_LICENSE("GPL v2");