builtin-kwork.c 46 KB

12345678910111213141516171819202122232425262728293031323334353637383940414243444546474849505152535455565758596061626364656667686970717273747576777879808182838485868788899091929394959697989910010110210310410510610710810911011111211311411511611711811912012112212312412512612712812913013113213313413513613713813914014114214314414514614714814915015115215315415515615715815916016116216316416516616716816917017117217317417517617717817918018118218318418518618718818919019119219319419519619719819920020120220320420520620720820921021121221321421521621721821922022122222322422522622722822923023123223323423523623723823924024124224324424524624724824925025125225325425525625725825926026126226326426526626726826927027127227327427527627727827928028128228328428528628728828929029129229329429529629729829930030130230330430530630730830931031131231331431531631731831932032132232332432532632732832933033133233333433533633733833934034134234334434534634734834935035135235335435535635735835936036136236336436536636736836937037137237337437537637737837938038138238338438538638738838939039139239339439539639739839940040140240340440540640740840941041141241341441541641741841942042142242342442542642742842943043143243343443543643743843944044144244344444544644744844945045145245345445545645745845946046146246346446546646746846947047147247347447547647747847948048148248348448548648748848949049149249349449549649749849950050150250350450550650750850951051151251351451551651751851952052152252352452552652752852953053153253353453553653753853954054154254354454554654754854955055155255355455555655755855956056156256356456556656756856957057157257357457557657757857958058158258358458558658758858959059159259359459559659759859960060160260360460560660760860961061161261361461561661761861962062162262362462562662762862963063163263363463563663763863964064164264364464564664764864965065165265365465565665765865966066166266366466566666766866967067167267367467567667767867968068168268368468568668768868969069169269369469569669769869970070170270370470570670770870971071171271371471571671771871972072172272372472572672772872973073173273373473573673773873974074174274374474574674774874975075175275375475575675775875976076176276376476576676776876977077177277377477577677777877978078178278378478578678778878979079179279379479579679779879980080180280380480580680780880981081181281381481581681781881982082182282382482582682782882983083183283383483583683783883984084184284384484584684784884985085185285385485585685785885986086186286386486586686786886987087187287387487587687787887988088188288388488588688788888989089189289389489589689789889990090190290390490590690790890991091191291391491591691791891992092192292392492592692792892993093193293393493593693793893994094194294394494594694794894995095195295395495595695795895996096196296396496596696796896997097197297397497597697797897998098198298398498598698798898999099199299399499599699799899910001001100210031004100510061007100810091010101110121013101410151016101710181019102010211022102310241025102610271028102910301031103210331034103510361037103810391040104110421043104410451046104710481049105010511052105310541055105610571058105910601061106210631064106510661067106810691070107110721073107410751076107710781079108010811082108310841085108610871088108910901091109210931094109510961097109810991100110111021103110411051106110711081109111011111112111311141115111611171118111911201121112211231124112511261127112811291130113111321133113411351136113711381139114011411142114311441145114611471148114911501151115211531154115511561157115811591160116111621163116411651166116711681169117011711172117311741175117611771178117911801181118211831184118511861187118811891190119111921193119411951196119711981199120012011202120312041205120612071208120912101211121212131214121512161217121812191220122112221223122412251226122712281229123012311232123312341235123612371238123912401241124212431244124512461247124812491250125112521253125412551256125712581259126012611262126312641265126612671268126912701271127212731274127512761277127812791280128112821283128412851286128712881289129012911292129312941295129612971298129913001301130213031304130513061307130813091310131113121313131413151316131713181319132013211322132313241325132613271328132913301331133213331334133513361337133813391340134113421343134413451346134713481349135013511352135313541355135613571358135913601361136213631364136513661367136813691370137113721373137413751376137713781379138013811382138313841385138613871388138913901391139213931394139513961397139813991400140114021403140414051406140714081409141014111412141314141415141614171418141914201421142214231424142514261427142814291430143114321433143414351436143714381439144014411442144314441445144614471448144914501451145214531454145514561457145814591460146114621463146414651466146714681469147014711472147314741475147614771478147914801481148214831484148514861487148814891490149114921493149414951496149714981499150015011502150315041505150615071508150915101511151215131514151515161517151815191520152115221523152415251526152715281529153015311532153315341535153615371538153915401541154215431544154515461547154815491550155115521553155415551556155715581559156015611562156315641565156615671568156915701571157215731574157515761577157815791580158115821583158415851586158715881589159015911592159315941595159615971598159916001601160216031604160516061607160816091610161116121613161416151616161716181619162016211622162316241625162616271628162916301631163216331634163516361637163816391640164116421643164416451646164716481649165016511652165316541655165616571658165916601661166216631664166516661667166816691670167116721673167416751676167716781679168016811682168316841685168616871688168916901691169216931694169516961697169816991700170117021703170417051706170717081709171017111712171317141715171617171718171917201721172217231724172517261727172817291730173117321733173417351736173717381739174017411742174317441745174617471748174917501751175217531754175517561757175817591760176117621763176417651766176717681769177017711772177317741775177617771778177917801781178217831784178517861787178817891790179117921793179417951796179717981799180018011802180318041805180618071808180918101811181218131814181518161817181818191820182118221823182418251826182718281829183018311832183318341835
  1. // SPDX-License-Identifier: GPL-2.0
  2. /*
  3. * builtin-kwork.c
  4. *
  5. * Copyright (c) 2022 Huawei Inc, Yang Jihong <[email protected]>
  6. */
  7. #include "builtin.h"
  8. #include "util/data.h"
  9. #include "util/kwork.h"
  10. #include "util/debug.h"
  11. #include "util/symbol.h"
  12. #include "util/thread.h"
  13. #include "util/string2.h"
  14. #include "util/callchain.h"
  15. #include "util/evsel_fprintf.h"
  16. #include <subcmd/pager.h>
  17. #include <subcmd/parse-options.h>
  18. #include <errno.h>
  19. #include <inttypes.h>
  20. #include <linux/err.h>
  21. #include <linux/time64.h>
  22. #include <linux/zalloc.h>
  23. /*
  24. * report header elements width
  25. */
  26. #define PRINT_CPU_WIDTH 4
  27. #define PRINT_COUNT_WIDTH 9
  28. #define PRINT_RUNTIME_WIDTH 10
  29. #define PRINT_LATENCY_WIDTH 10
  30. #define PRINT_TIMESTAMP_WIDTH 17
  31. #define PRINT_KWORK_NAME_WIDTH 30
  32. #define RPINT_DECIMAL_WIDTH 3
  33. #define PRINT_BRACKETPAIR_WIDTH 2
  34. #define PRINT_TIME_UNIT_SEC_WIDTH 2
  35. #define PRINT_TIME_UNIT_MESC_WIDTH 3
  36. #define PRINT_RUNTIME_HEADER_WIDTH (PRINT_RUNTIME_WIDTH + PRINT_TIME_UNIT_MESC_WIDTH)
  37. #define PRINT_LATENCY_HEADER_WIDTH (PRINT_LATENCY_WIDTH + PRINT_TIME_UNIT_MESC_WIDTH)
  38. #define PRINT_TIMEHIST_CPU_WIDTH (PRINT_CPU_WIDTH + PRINT_BRACKETPAIR_WIDTH)
  39. #define PRINT_TIMESTAMP_HEADER_WIDTH (PRINT_TIMESTAMP_WIDTH + PRINT_TIME_UNIT_SEC_WIDTH)
  40. struct sort_dimension {
  41. const char *name;
  42. int (*cmp)(struct kwork_work *l, struct kwork_work *r);
  43. struct list_head list;
  44. };
  45. static int id_cmp(struct kwork_work *l, struct kwork_work *r)
  46. {
  47. if (l->cpu > r->cpu)
  48. return 1;
  49. if (l->cpu < r->cpu)
  50. return -1;
  51. if (l->id > r->id)
  52. return 1;
  53. if (l->id < r->id)
  54. return -1;
  55. return 0;
  56. }
  57. static int count_cmp(struct kwork_work *l, struct kwork_work *r)
  58. {
  59. if (l->nr_atoms > r->nr_atoms)
  60. return 1;
  61. if (l->nr_atoms < r->nr_atoms)
  62. return -1;
  63. return 0;
  64. }
  65. static int runtime_cmp(struct kwork_work *l, struct kwork_work *r)
  66. {
  67. if (l->total_runtime > r->total_runtime)
  68. return 1;
  69. if (l->total_runtime < r->total_runtime)
  70. return -1;
  71. return 0;
  72. }
  73. static int max_runtime_cmp(struct kwork_work *l, struct kwork_work *r)
  74. {
  75. if (l->max_runtime > r->max_runtime)
  76. return 1;
  77. if (l->max_runtime < r->max_runtime)
  78. return -1;
  79. return 0;
  80. }
  81. static int avg_latency_cmp(struct kwork_work *l, struct kwork_work *r)
  82. {
  83. u64 avgl, avgr;
  84. if (!r->nr_atoms)
  85. return 1;
  86. if (!l->nr_atoms)
  87. return -1;
  88. avgl = l->total_latency / l->nr_atoms;
  89. avgr = r->total_latency / r->nr_atoms;
  90. if (avgl > avgr)
  91. return 1;
  92. if (avgl < avgr)
  93. return -1;
  94. return 0;
  95. }
  96. static int max_latency_cmp(struct kwork_work *l, struct kwork_work *r)
  97. {
  98. if (l->max_latency > r->max_latency)
  99. return 1;
  100. if (l->max_latency < r->max_latency)
  101. return -1;
  102. return 0;
  103. }
  104. static int sort_dimension__add(struct perf_kwork *kwork __maybe_unused,
  105. const char *tok, struct list_head *list)
  106. {
  107. size_t i;
  108. static struct sort_dimension max_sort_dimension = {
  109. .name = "max",
  110. .cmp = max_runtime_cmp,
  111. };
  112. static struct sort_dimension id_sort_dimension = {
  113. .name = "id",
  114. .cmp = id_cmp,
  115. };
  116. static struct sort_dimension runtime_sort_dimension = {
  117. .name = "runtime",
  118. .cmp = runtime_cmp,
  119. };
  120. static struct sort_dimension count_sort_dimension = {
  121. .name = "count",
  122. .cmp = count_cmp,
  123. };
  124. static struct sort_dimension avg_sort_dimension = {
  125. .name = "avg",
  126. .cmp = avg_latency_cmp,
  127. };
  128. struct sort_dimension *available_sorts[] = {
  129. &id_sort_dimension,
  130. &max_sort_dimension,
  131. &count_sort_dimension,
  132. &runtime_sort_dimension,
  133. &avg_sort_dimension,
  134. };
  135. if (kwork->report == KWORK_REPORT_LATENCY)
  136. max_sort_dimension.cmp = max_latency_cmp;
  137. for (i = 0; i < ARRAY_SIZE(available_sorts); i++) {
  138. if (!strcmp(available_sorts[i]->name, tok)) {
  139. list_add_tail(&available_sorts[i]->list, list);
  140. return 0;
  141. }
  142. }
  143. return -1;
  144. }
  145. static void setup_sorting(struct perf_kwork *kwork,
  146. const struct option *options,
  147. const char * const usage_msg[])
  148. {
  149. char *tmp, *tok, *str = strdup(kwork->sort_order);
  150. for (tok = strtok_r(str, ", ", &tmp);
  151. tok; tok = strtok_r(NULL, ", ", &tmp)) {
  152. if (sort_dimension__add(kwork, tok, &kwork->sort_list) < 0)
  153. usage_with_options_msg(usage_msg, options,
  154. "Unknown --sort key: `%s'", tok);
  155. }
  156. pr_debug("Sort order: %s\n", kwork->sort_order);
  157. free(str);
  158. }
  159. static struct kwork_atom *atom_new(struct perf_kwork *kwork,
  160. struct perf_sample *sample)
  161. {
  162. unsigned long i;
  163. struct kwork_atom_page *page;
  164. struct kwork_atom *atom = NULL;
  165. list_for_each_entry(page, &kwork->atom_page_list, list) {
  166. if (!bitmap_full(page->bitmap, NR_ATOM_PER_PAGE)) {
  167. i = find_first_zero_bit(page->bitmap, NR_ATOM_PER_PAGE);
  168. BUG_ON(i >= NR_ATOM_PER_PAGE);
  169. atom = &page->atoms[i];
  170. goto found_atom;
  171. }
  172. }
  173. /*
  174. * new page
  175. */
  176. page = zalloc(sizeof(*page));
  177. if (page == NULL) {
  178. pr_err("Failed to zalloc kwork atom page\n");
  179. return NULL;
  180. }
  181. i = 0;
  182. atom = &page->atoms[0];
  183. list_add_tail(&page->list, &kwork->atom_page_list);
  184. found_atom:
  185. set_bit(i, page->bitmap);
  186. atom->time = sample->time;
  187. atom->prev = NULL;
  188. atom->page_addr = page;
  189. atom->bit_inpage = i;
  190. return atom;
  191. }
  192. static void atom_free(struct kwork_atom *atom)
  193. {
  194. if (atom->prev != NULL)
  195. atom_free(atom->prev);
  196. clear_bit(atom->bit_inpage,
  197. ((struct kwork_atom_page *)atom->page_addr)->bitmap);
  198. }
  199. static void atom_del(struct kwork_atom *atom)
  200. {
  201. list_del(&atom->list);
  202. atom_free(atom);
  203. }
  204. static int work_cmp(struct list_head *list,
  205. struct kwork_work *l, struct kwork_work *r)
  206. {
  207. int ret = 0;
  208. struct sort_dimension *sort;
  209. BUG_ON(list_empty(list));
  210. list_for_each_entry(sort, list, list) {
  211. ret = sort->cmp(l, r);
  212. if (ret)
  213. return ret;
  214. }
  215. return ret;
  216. }
  217. static struct kwork_work *work_search(struct rb_root_cached *root,
  218. struct kwork_work *key,
  219. struct list_head *sort_list)
  220. {
  221. int cmp;
  222. struct kwork_work *work;
  223. struct rb_node *node = root->rb_root.rb_node;
  224. while (node) {
  225. work = container_of(node, struct kwork_work, node);
  226. cmp = work_cmp(sort_list, key, work);
  227. if (cmp > 0)
  228. node = node->rb_left;
  229. else if (cmp < 0)
  230. node = node->rb_right;
  231. else {
  232. if (work->name == NULL)
  233. work->name = key->name;
  234. return work;
  235. }
  236. }
  237. return NULL;
  238. }
  239. static void work_insert(struct rb_root_cached *root,
  240. struct kwork_work *key, struct list_head *sort_list)
  241. {
  242. int cmp;
  243. bool leftmost = true;
  244. struct kwork_work *cur;
  245. struct rb_node **new = &(root->rb_root.rb_node), *parent = NULL;
  246. while (*new) {
  247. cur = container_of(*new, struct kwork_work, node);
  248. parent = *new;
  249. cmp = work_cmp(sort_list, key, cur);
  250. if (cmp > 0)
  251. new = &((*new)->rb_left);
  252. else {
  253. new = &((*new)->rb_right);
  254. leftmost = false;
  255. }
  256. }
  257. rb_link_node(&key->node, parent, new);
  258. rb_insert_color_cached(&key->node, root, leftmost);
  259. }
  260. static struct kwork_work *work_new(struct kwork_work *key)
  261. {
  262. int i;
  263. struct kwork_work *work = zalloc(sizeof(*work));
  264. if (work == NULL) {
  265. pr_err("Failed to zalloc kwork work\n");
  266. return NULL;
  267. }
  268. for (i = 0; i < KWORK_TRACE_MAX; i++)
  269. INIT_LIST_HEAD(&work->atom_list[i]);
  270. work->id = key->id;
  271. work->cpu = key->cpu;
  272. work->name = key->name;
  273. work->class = key->class;
  274. return work;
  275. }
  276. static struct kwork_work *work_findnew(struct rb_root_cached *root,
  277. struct kwork_work *key,
  278. struct list_head *sort_list)
  279. {
  280. struct kwork_work *work = work_search(root, key, sort_list);
  281. if (work != NULL)
  282. return work;
  283. work = work_new(key);
  284. if (work)
  285. work_insert(root, work, sort_list);
  286. return work;
  287. }
  288. static void profile_update_timespan(struct perf_kwork *kwork,
  289. struct perf_sample *sample)
  290. {
  291. if (!kwork->summary)
  292. return;
  293. if ((kwork->timestart == 0) || (kwork->timestart > sample->time))
  294. kwork->timestart = sample->time;
  295. if (kwork->timeend < sample->time)
  296. kwork->timeend = sample->time;
  297. }
  298. static bool profile_event_match(struct perf_kwork *kwork,
  299. struct kwork_work *work,
  300. struct perf_sample *sample)
  301. {
  302. int cpu = work->cpu;
  303. u64 time = sample->time;
  304. struct perf_time_interval *ptime = &kwork->ptime;
  305. if ((kwork->cpu_list != NULL) && !test_bit(cpu, kwork->cpu_bitmap))
  306. return false;
  307. if (((ptime->start != 0) && (ptime->start > time)) ||
  308. ((ptime->end != 0) && (ptime->end < time)))
  309. return false;
  310. if ((kwork->profile_name != NULL) &&
  311. (work->name != NULL) &&
  312. (strcmp(work->name, kwork->profile_name) != 0))
  313. return false;
  314. profile_update_timespan(kwork, sample);
  315. return true;
  316. }
  317. static int work_push_atom(struct perf_kwork *kwork,
  318. struct kwork_class *class,
  319. enum kwork_trace_type src_type,
  320. enum kwork_trace_type dst_type,
  321. struct evsel *evsel,
  322. struct perf_sample *sample,
  323. struct machine *machine,
  324. struct kwork_work **ret_work)
  325. {
  326. struct kwork_atom *atom, *dst_atom;
  327. struct kwork_work *work, key;
  328. BUG_ON(class->work_init == NULL);
  329. class->work_init(class, &key, evsel, sample, machine);
  330. atom = atom_new(kwork, sample);
  331. if (atom == NULL)
  332. return -1;
  333. work = work_findnew(&class->work_root, &key, &kwork->cmp_id);
  334. if (work == NULL) {
  335. atom_free(atom);
  336. return -1;
  337. }
  338. if (!profile_event_match(kwork, work, sample)) {
  339. atom_free(atom);
  340. return 0;
  341. }
  342. if (dst_type < KWORK_TRACE_MAX) {
  343. dst_atom = list_last_entry_or_null(&work->atom_list[dst_type],
  344. struct kwork_atom, list);
  345. if (dst_atom != NULL) {
  346. atom->prev = dst_atom;
  347. list_del(&dst_atom->list);
  348. }
  349. }
  350. if (ret_work != NULL)
  351. *ret_work = work;
  352. list_add_tail(&atom->list, &work->atom_list[src_type]);
  353. return 0;
  354. }
  355. static struct kwork_atom *work_pop_atom(struct perf_kwork *kwork,
  356. struct kwork_class *class,
  357. enum kwork_trace_type src_type,
  358. enum kwork_trace_type dst_type,
  359. struct evsel *evsel,
  360. struct perf_sample *sample,
  361. struct machine *machine,
  362. struct kwork_work **ret_work)
  363. {
  364. struct kwork_atom *atom, *src_atom;
  365. struct kwork_work *work, key;
  366. BUG_ON(class->work_init == NULL);
  367. class->work_init(class, &key, evsel, sample, machine);
  368. work = work_findnew(&class->work_root, &key, &kwork->cmp_id);
  369. if (ret_work != NULL)
  370. *ret_work = work;
  371. if (work == NULL)
  372. return NULL;
  373. if (!profile_event_match(kwork, work, sample))
  374. return NULL;
  375. atom = list_last_entry_or_null(&work->atom_list[dst_type],
  376. struct kwork_atom, list);
  377. if (atom != NULL)
  378. return atom;
  379. src_atom = atom_new(kwork, sample);
  380. if (src_atom != NULL)
  381. list_add_tail(&src_atom->list, &work->atom_list[src_type]);
  382. else {
  383. if (ret_work != NULL)
  384. *ret_work = NULL;
  385. }
  386. return NULL;
  387. }
  388. static void report_update_exit_event(struct kwork_work *work,
  389. struct kwork_atom *atom,
  390. struct perf_sample *sample)
  391. {
  392. u64 delta;
  393. u64 exit_time = sample->time;
  394. u64 entry_time = atom->time;
  395. if ((entry_time != 0) && (exit_time >= entry_time)) {
  396. delta = exit_time - entry_time;
  397. if ((delta > work->max_runtime) ||
  398. (work->max_runtime == 0)) {
  399. work->max_runtime = delta;
  400. work->max_runtime_start = entry_time;
  401. work->max_runtime_end = exit_time;
  402. }
  403. work->total_runtime += delta;
  404. work->nr_atoms++;
  405. }
  406. }
  407. static int report_entry_event(struct perf_kwork *kwork,
  408. struct kwork_class *class,
  409. struct evsel *evsel,
  410. struct perf_sample *sample,
  411. struct machine *machine)
  412. {
  413. return work_push_atom(kwork, class, KWORK_TRACE_ENTRY,
  414. KWORK_TRACE_MAX, evsel, sample,
  415. machine, NULL);
  416. }
  417. static int report_exit_event(struct perf_kwork *kwork,
  418. struct kwork_class *class,
  419. struct evsel *evsel,
  420. struct perf_sample *sample,
  421. struct machine *machine)
  422. {
  423. struct kwork_atom *atom = NULL;
  424. struct kwork_work *work = NULL;
  425. atom = work_pop_atom(kwork, class, KWORK_TRACE_EXIT,
  426. KWORK_TRACE_ENTRY, evsel, sample,
  427. machine, &work);
  428. if (work == NULL)
  429. return -1;
  430. if (atom != NULL) {
  431. report_update_exit_event(work, atom, sample);
  432. atom_del(atom);
  433. }
  434. return 0;
  435. }
  436. static void latency_update_entry_event(struct kwork_work *work,
  437. struct kwork_atom *atom,
  438. struct perf_sample *sample)
  439. {
  440. u64 delta;
  441. u64 entry_time = sample->time;
  442. u64 raise_time = atom->time;
  443. if ((raise_time != 0) && (entry_time >= raise_time)) {
  444. delta = entry_time - raise_time;
  445. if ((delta > work->max_latency) ||
  446. (work->max_latency == 0)) {
  447. work->max_latency = delta;
  448. work->max_latency_start = raise_time;
  449. work->max_latency_end = entry_time;
  450. }
  451. work->total_latency += delta;
  452. work->nr_atoms++;
  453. }
  454. }
  455. static int latency_raise_event(struct perf_kwork *kwork,
  456. struct kwork_class *class,
  457. struct evsel *evsel,
  458. struct perf_sample *sample,
  459. struct machine *machine)
  460. {
  461. return work_push_atom(kwork, class, KWORK_TRACE_RAISE,
  462. KWORK_TRACE_MAX, evsel, sample,
  463. machine, NULL);
  464. }
  465. static int latency_entry_event(struct perf_kwork *kwork,
  466. struct kwork_class *class,
  467. struct evsel *evsel,
  468. struct perf_sample *sample,
  469. struct machine *machine)
  470. {
  471. struct kwork_atom *atom = NULL;
  472. struct kwork_work *work = NULL;
  473. atom = work_pop_atom(kwork, class, KWORK_TRACE_ENTRY,
  474. KWORK_TRACE_RAISE, evsel, sample,
  475. machine, &work);
  476. if (work == NULL)
  477. return -1;
  478. if (atom != NULL) {
  479. latency_update_entry_event(work, atom, sample);
  480. atom_del(atom);
  481. }
  482. return 0;
  483. }
  484. static void timehist_save_callchain(struct perf_kwork *kwork,
  485. struct perf_sample *sample,
  486. struct evsel *evsel,
  487. struct machine *machine)
  488. {
  489. struct symbol *sym;
  490. struct thread *thread;
  491. struct callchain_cursor_node *node;
  492. struct callchain_cursor *cursor = &callchain_cursor;
  493. if (!kwork->show_callchain || sample->callchain == NULL)
  494. return;
  495. /* want main thread for process - has maps */
  496. thread = machine__findnew_thread(machine, sample->pid, sample->pid);
  497. if (thread == NULL) {
  498. pr_debug("Failed to get thread for pid %d\n", sample->pid);
  499. return;
  500. }
  501. if (thread__resolve_callchain(thread, cursor, evsel, sample,
  502. NULL, NULL, kwork->max_stack + 2) != 0) {
  503. pr_debug("Failed to resolve callchain, skipping\n");
  504. goto out_put;
  505. }
  506. callchain_cursor_commit(cursor);
  507. while (true) {
  508. node = callchain_cursor_current(cursor);
  509. if (node == NULL)
  510. break;
  511. sym = node->ms.sym;
  512. if (sym) {
  513. if (!strcmp(sym->name, "__softirqentry_text_start") ||
  514. !strcmp(sym->name, "__do_softirq"))
  515. sym->ignore = 1;
  516. }
  517. callchain_cursor_advance(cursor);
  518. }
  519. out_put:
  520. thread__put(thread);
  521. }
  522. static void timehist_print_event(struct perf_kwork *kwork,
  523. struct kwork_work *work,
  524. struct kwork_atom *atom,
  525. struct perf_sample *sample,
  526. struct addr_location *al)
  527. {
  528. char entrytime[32], exittime[32];
  529. char kwork_name[PRINT_KWORK_NAME_WIDTH];
  530. /*
  531. * runtime start
  532. */
  533. timestamp__scnprintf_usec(atom->time,
  534. entrytime, sizeof(entrytime));
  535. printf(" %*s ", PRINT_TIMESTAMP_WIDTH, entrytime);
  536. /*
  537. * runtime end
  538. */
  539. timestamp__scnprintf_usec(sample->time,
  540. exittime, sizeof(exittime));
  541. printf(" %*s ", PRINT_TIMESTAMP_WIDTH, exittime);
  542. /*
  543. * cpu
  544. */
  545. printf(" [%0*d] ", PRINT_CPU_WIDTH, work->cpu);
  546. /*
  547. * kwork name
  548. */
  549. if (work->class && work->class->work_name) {
  550. work->class->work_name(work, kwork_name,
  551. PRINT_KWORK_NAME_WIDTH);
  552. printf(" %-*s ", PRINT_KWORK_NAME_WIDTH, kwork_name);
  553. } else
  554. printf(" %-*s ", PRINT_KWORK_NAME_WIDTH, "");
  555. /*
  556. *runtime
  557. */
  558. printf(" %*.*f ",
  559. PRINT_RUNTIME_WIDTH, RPINT_DECIMAL_WIDTH,
  560. (double)(sample->time - atom->time) / NSEC_PER_MSEC);
  561. /*
  562. * delaytime
  563. */
  564. if (atom->prev != NULL)
  565. printf(" %*.*f ", PRINT_LATENCY_WIDTH, RPINT_DECIMAL_WIDTH,
  566. (double)(atom->time - atom->prev->time) / NSEC_PER_MSEC);
  567. else
  568. printf(" %*s ", PRINT_LATENCY_WIDTH, " ");
  569. /*
  570. * callchain
  571. */
  572. if (kwork->show_callchain) {
  573. printf(" ");
  574. sample__fprintf_sym(sample, al, 0,
  575. EVSEL__PRINT_SYM | EVSEL__PRINT_ONELINE |
  576. EVSEL__PRINT_CALLCHAIN_ARROW |
  577. EVSEL__PRINT_SKIP_IGNORED,
  578. &callchain_cursor, symbol_conf.bt_stop_list,
  579. stdout);
  580. }
  581. printf("\n");
  582. }
  583. static int timehist_raise_event(struct perf_kwork *kwork,
  584. struct kwork_class *class,
  585. struct evsel *evsel,
  586. struct perf_sample *sample,
  587. struct machine *machine)
  588. {
  589. return work_push_atom(kwork, class, KWORK_TRACE_RAISE,
  590. KWORK_TRACE_MAX, evsel, sample,
  591. machine, NULL);
  592. }
  593. static int timehist_entry_event(struct perf_kwork *kwork,
  594. struct kwork_class *class,
  595. struct evsel *evsel,
  596. struct perf_sample *sample,
  597. struct machine *machine)
  598. {
  599. int ret;
  600. struct kwork_work *work = NULL;
  601. ret = work_push_atom(kwork, class, KWORK_TRACE_ENTRY,
  602. KWORK_TRACE_RAISE, evsel, sample,
  603. machine, &work);
  604. if (ret)
  605. return ret;
  606. if (work != NULL)
  607. timehist_save_callchain(kwork, sample, evsel, machine);
  608. return 0;
  609. }
  610. static int timehist_exit_event(struct perf_kwork *kwork,
  611. struct kwork_class *class,
  612. struct evsel *evsel,
  613. struct perf_sample *sample,
  614. struct machine *machine)
  615. {
  616. struct kwork_atom *atom = NULL;
  617. struct kwork_work *work = NULL;
  618. struct addr_location al;
  619. if (machine__resolve(machine, &al, sample) < 0) {
  620. pr_debug("Problem processing event, skipping it\n");
  621. return -1;
  622. }
  623. atom = work_pop_atom(kwork, class, KWORK_TRACE_EXIT,
  624. KWORK_TRACE_ENTRY, evsel, sample,
  625. machine, &work);
  626. if (work == NULL)
  627. return -1;
  628. if (atom != NULL) {
  629. work->nr_atoms++;
  630. timehist_print_event(kwork, work, atom, sample, &al);
  631. atom_del(atom);
  632. }
  633. return 0;
  634. }
  635. static struct kwork_class kwork_irq;
  636. static int process_irq_handler_entry_event(struct perf_tool *tool,
  637. struct evsel *evsel,
  638. struct perf_sample *sample,
  639. struct machine *machine)
  640. {
  641. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  642. if (kwork->tp_handler->entry_event)
  643. return kwork->tp_handler->entry_event(kwork, &kwork_irq,
  644. evsel, sample, machine);
  645. return 0;
  646. }
  647. static int process_irq_handler_exit_event(struct perf_tool *tool,
  648. struct evsel *evsel,
  649. struct perf_sample *sample,
  650. struct machine *machine)
  651. {
  652. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  653. if (kwork->tp_handler->exit_event)
  654. return kwork->tp_handler->exit_event(kwork, &kwork_irq,
  655. evsel, sample, machine);
  656. return 0;
  657. }
  658. const struct evsel_str_handler irq_tp_handlers[] = {
  659. { "irq:irq_handler_entry", process_irq_handler_entry_event, },
  660. { "irq:irq_handler_exit", process_irq_handler_exit_event, },
  661. };
  662. static int irq_class_init(struct kwork_class *class,
  663. struct perf_session *session)
  664. {
  665. if (perf_session__set_tracepoints_handlers(session, irq_tp_handlers)) {
  666. pr_err("Failed to set irq tracepoints handlers\n");
  667. return -1;
  668. }
  669. class->work_root = RB_ROOT_CACHED;
  670. return 0;
  671. }
  672. static void irq_work_init(struct kwork_class *class,
  673. struct kwork_work *work,
  674. struct evsel *evsel,
  675. struct perf_sample *sample,
  676. struct machine *machine __maybe_unused)
  677. {
  678. work->class = class;
  679. work->cpu = sample->cpu;
  680. work->id = evsel__intval(evsel, sample, "irq");
  681. work->name = evsel__strval(evsel, sample, "name");
  682. }
  683. static void irq_work_name(struct kwork_work *work, char *buf, int len)
  684. {
  685. snprintf(buf, len, "%s:%" PRIu64 "", work->name, work->id);
  686. }
  687. static struct kwork_class kwork_irq = {
  688. .name = "irq",
  689. .type = KWORK_CLASS_IRQ,
  690. .nr_tracepoints = 2,
  691. .tp_handlers = irq_tp_handlers,
  692. .class_init = irq_class_init,
  693. .work_init = irq_work_init,
  694. .work_name = irq_work_name,
  695. };
  696. static struct kwork_class kwork_softirq;
  697. static int process_softirq_raise_event(struct perf_tool *tool,
  698. struct evsel *evsel,
  699. struct perf_sample *sample,
  700. struct machine *machine)
  701. {
  702. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  703. if (kwork->tp_handler->raise_event)
  704. return kwork->tp_handler->raise_event(kwork, &kwork_softirq,
  705. evsel, sample, machine);
  706. return 0;
  707. }
  708. static int process_softirq_entry_event(struct perf_tool *tool,
  709. struct evsel *evsel,
  710. struct perf_sample *sample,
  711. struct machine *machine)
  712. {
  713. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  714. if (kwork->tp_handler->entry_event)
  715. return kwork->tp_handler->entry_event(kwork, &kwork_softirq,
  716. evsel, sample, machine);
  717. return 0;
  718. }
  719. static int process_softirq_exit_event(struct perf_tool *tool,
  720. struct evsel *evsel,
  721. struct perf_sample *sample,
  722. struct machine *machine)
  723. {
  724. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  725. if (kwork->tp_handler->exit_event)
  726. return kwork->tp_handler->exit_event(kwork, &kwork_softirq,
  727. evsel, sample, machine);
  728. return 0;
  729. }
  730. const struct evsel_str_handler softirq_tp_handlers[] = {
  731. { "irq:softirq_raise", process_softirq_raise_event, },
  732. { "irq:softirq_entry", process_softirq_entry_event, },
  733. { "irq:softirq_exit", process_softirq_exit_event, },
  734. };
  735. static int softirq_class_init(struct kwork_class *class,
  736. struct perf_session *session)
  737. {
  738. if (perf_session__set_tracepoints_handlers(session,
  739. softirq_tp_handlers)) {
  740. pr_err("Failed to set softirq tracepoints handlers\n");
  741. return -1;
  742. }
  743. class->work_root = RB_ROOT_CACHED;
  744. return 0;
  745. }
  746. static char *evsel__softirq_name(struct evsel *evsel, u64 num)
  747. {
  748. char *name = NULL;
  749. bool found = false;
  750. struct tep_print_flag_sym *sym = NULL;
  751. struct tep_print_arg *args = evsel->tp_format->print_fmt.args;
  752. if ((args == NULL) || (args->next == NULL))
  753. return NULL;
  754. /* skip softirq field: "REC->vec" */
  755. for (sym = args->next->symbol.symbols; sym != NULL; sym = sym->next) {
  756. if ((eval_flag(sym->value) == (unsigned long long)num) &&
  757. (strlen(sym->str) != 0)) {
  758. found = true;
  759. break;
  760. }
  761. }
  762. if (!found)
  763. return NULL;
  764. name = strdup(sym->str);
  765. if (name == NULL) {
  766. pr_err("Failed to copy symbol name\n");
  767. return NULL;
  768. }
  769. return name;
  770. }
  771. static void softirq_work_init(struct kwork_class *class,
  772. struct kwork_work *work,
  773. struct evsel *evsel,
  774. struct perf_sample *sample,
  775. struct machine *machine __maybe_unused)
  776. {
  777. u64 num = evsel__intval(evsel, sample, "vec");
  778. work->id = num;
  779. work->class = class;
  780. work->cpu = sample->cpu;
  781. work->name = evsel__softirq_name(evsel, num);
  782. }
  783. static void softirq_work_name(struct kwork_work *work, char *buf, int len)
  784. {
  785. snprintf(buf, len, "(s)%s:%" PRIu64 "", work->name, work->id);
  786. }
  787. static struct kwork_class kwork_softirq = {
  788. .name = "softirq",
  789. .type = KWORK_CLASS_SOFTIRQ,
  790. .nr_tracepoints = 3,
  791. .tp_handlers = softirq_tp_handlers,
  792. .class_init = softirq_class_init,
  793. .work_init = softirq_work_init,
  794. .work_name = softirq_work_name,
  795. };
  796. static struct kwork_class kwork_workqueue;
  797. static int process_workqueue_activate_work_event(struct perf_tool *tool,
  798. struct evsel *evsel,
  799. struct perf_sample *sample,
  800. struct machine *machine)
  801. {
  802. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  803. if (kwork->tp_handler->raise_event)
  804. return kwork->tp_handler->raise_event(kwork, &kwork_workqueue,
  805. evsel, sample, machine);
  806. return 0;
  807. }
  808. static int process_workqueue_execute_start_event(struct perf_tool *tool,
  809. struct evsel *evsel,
  810. struct perf_sample *sample,
  811. struct machine *machine)
  812. {
  813. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  814. if (kwork->tp_handler->entry_event)
  815. return kwork->tp_handler->entry_event(kwork, &kwork_workqueue,
  816. evsel, sample, machine);
  817. return 0;
  818. }
  819. static int process_workqueue_execute_end_event(struct perf_tool *tool,
  820. struct evsel *evsel,
  821. struct perf_sample *sample,
  822. struct machine *machine)
  823. {
  824. struct perf_kwork *kwork = container_of(tool, struct perf_kwork, tool);
  825. if (kwork->tp_handler->exit_event)
  826. return kwork->tp_handler->exit_event(kwork, &kwork_workqueue,
  827. evsel, sample, machine);
  828. return 0;
  829. }
  830. const struct evsel_str_handler workqueue_tp_handlers[] = {
  831. { "workqueue:workqueue_activate_work", process_workqueue_activate_work_event, },
  832. { "workqueue:workqueue_execute_start", process_workqueue_execute_start_event, },
  833. { "workqueue:workqueue_execute_end", process_workqueue_execute_end_event, },
  834. };
  835. static int workqueue_class_init(struct kwork_class *class,
  836. struct perf_session *session)
  837. {
  838. if (perf_session__set_tracepoints_handlers(session,
  839. workqueue_tp_handlers)) {
  840. pr_err("Failed to set workqueue tracepoints handlers\n");
  841. return -1;
  842. }
  843. class->work_root = RB_ROOT_CACHED;
  844. return 0;
  845. }
  846. static void workqueue_work_init(struct kwork_class *class,
  847. struct kwork_work *work,
  848. struct evsel *evsel,
  849. struct perf_sample *sample,
  850. struct machine *machine)
  851. {
  852. char *modp = NULL;
  853. unsigned long long function_addr = evsel__intval(evsel,
  854. sample, "function");
  855. work->class = class;
  856. work->cpu = sample->cpu;
  857. work->id = evsel__intval(evsel, sample, "work");
  858. work->name = function_addr == 0 ? NULL :
  859. machine__resolve_kernel_addr(machine, &function_addr, &modp);
  860. }
  861. static void workqueue_work_name(struct kwork_work *work, char *buf, int len)
  862. {
  863. if (work->name != NULL)
  864. snprintf(buf, len, "(w)%s", work->name);
  865. else
  866. snprintf(buf, len, "(w)0x%" PRIx64, work->id);
  867. }
  868. static struct kwork_class kwork_workqueue = {
  869. .name = "workqueue",
  870. .type = KWORK_CLASS_WORKQUEUE,
  871. .nr_tracepoints = 3,
  872. .tp_handlers = workqueue_tp_handlers,
  873. .class_init = workqueue_class_init,
  874. .work_init = workqueue_work_init,
  875. .work_name = workqueue_work_name,
  876. };
  877. static struct kwork_class *kwork_class_supported_list[KWORK_CLASS_MAX] = {
  878. [KWORK_CLASS_IRQ] = &kwork_irq,
  879. [KWORK_CLASS_SOFTIRQ] = &kwork_softirq,
  880. [KWORK_CLASS_WORKQUEUE] = &kwork_workqueue,
  881. };
  882. static void print_separator(int len)
  883. {
  884. printf(" %.*s\n", len, graph_dotted_line);
  885. }
  886. static int report_print_work(struct perf_kwork *kwork, struct kwork_work *work)
  887. {
  888. int ret = 0;
  889. char kwork_name[PRINT_KWORK_NAME_WIDTH];
  890. char max_runtime_start[32], max_runtime_end[32];
  891. char max_latency_start[32], max_latency_end[32];
  892. printf(" ");
  893. /*
  894. * kwork name
  895. */
  896. if (work->class && work->class->work_name) {
  897. work->class->work_name(work, kwork_name,
  898. PRINT_KWORK_NAME_WIDTH);
  899. ret += printf(" %-*s |", PRINT_KWORK_NAME_WIDTH, kwork_name);
  900. } else {
  901. ret += printf(" %-*s |", PRINT_KWORK_NAME_WIDTH, "");
  902. }
  903. /*
  904. * cpu
  905. */
  906. ret += printf(" %0*d |", PRINT_CPU_WIDTH, work->cpu);
  907. /*
  908. * total runtime
  909. */
  910. if (kwork->report == KWORK_REPORT_RUNTIME) {
  911. ret += printf(" %*.*f ms |",
  912. PRINT_RUNTIME_WIDTH, RPINT_DECIMAL_WIDTH,
  913. (double)work->total_runtime / NSEC_PER_MSEC);
  914. } else if (kwork->report == KWORK_REPORT_LATENCY) { // avg delay
  915. ret += printf(" %*.*f ms |",
  916. PRINT_LATENCY_WIDTH, RPINT_DECIMAL_WIDTH,
  917. (double)work->total_latency /
  918. work->nr_atoms / NSEC_PER_MSEC);
  919. }
  920. /*
  921. * count
  922. */
  923. ret += printf(" %*" PRIu64 " |", PRINT_COUNT_WIDTH, work->nr_atoms);
  924. /*
  925. * max runtime, max runtime start, max runtime end
  926. */
  927. if (kwork->report == KWORK_REPORT_RUNTIME) {
  928. timestamp__scnprintf_usec(work->max_runtime_start,
  929. max_runtime_start,
  930. sizeof(max_runtime_start));
  931. timestamp__scnprintf_usec(work->max_runtime_end,
  932. max_runtime_end,
  933. sizeof(max_runtime_end));
  934. ret += printf(" %*.*f ms | %*s s | %*s s |",
  935. PRINT_RUNTIME_WIDTH, RPINT_DECIMAL_WIDTH,
  936. (double)work->max_runtime / NSEC_PER_MSEC,
  937. PRINT_TIMESTAMP_WIDTH, max_runtime_start,
  938. PRINT_TIMESTAMP_WIDTH, max_runtime_end);
  939. }
  940. /*
  941. * max delay, max delay start, max delay end
  942. */
  943. else if (kwork->report == KWORK_REPORT_LATENCY) {
  944. timestamp__scnprintf_usec(work->max_latency_start,
  945. max_latency_start,
  946. sizeof(max_latency_start));
  947. timestamp__scnprintf_usec(work->max_latency_end,
  948. max_latency_end,
  949. sizeof(max_latency_end));
  950. ret += printf(" %*.*f ms | %*s s | %*s s |",
  951. PRINT_LATENCY_WIDTH, RPINT_DECIMAL_WIDTH,
  952. (double)work->max_latency / NSEC_PER_MSEC,
  953. PRINT_TIMESTAMP_WIDTH, max_latency_start,
  954. PRINT_TIMESTAMP_WIDTH, max_latency_end);
  955. }
  956. printf("\n");
  957. return ret;
  958. }
  959. static int report_print_header(struct perf_kwork *kwork)
  960. {
  961. int ret;
  962. printf("\n ");
  963. ret = printf(" %-*s | %-*s |",
  964. PRINT_KWORK_NAME_WIDTH, "Kwork Name",
  965. PRINT_CPU_WIDTH, "Cpu");
  966. if (kwork->report == KWORK_REPORT_RUNTIME) {
  967. ret += printf(" %-*s |",
  968. PRINT_RUNTIME_HEADER_WIDTH, "Total Runtime");
  969. } else if (kwork->report == KWORK_REPORT_LATENCY) {
  970. ret += printf(" %-*s |",
  971. PRINT_LATENCY_HEADER_WIDTH, "Avg delay");
  972. }
  973. ret += printf(" %-*s |", PRINT_COUNT_WIDTH, "Count");
  974. if (kwork->report == KWORK_REPORT_RUNTIME) {
  975. ret += printf(" %-*s | %-*s | %-*s |",
  976. PRINT_RUNTIME_HEADER_WIDTH, "Max runtime",
  977. PRINT_TIMESTAMP_HEADER_WIDTH, "Max runtime start",
  978. PRINT_TIMESTAMP_HEADER_WIDTH, "Max runtime end");
  979. } else if (kwork->report == KWORK_REPORT_LATENCY) {
  980. ret += printf(" %-*s | %-*s | %-*s |",
  981. PRINT_LATENCY_HEADER_WIDTH, "Max delay",
  982. PRINT_TIMESTAMP_HEADER_WIDTH, "Max delay start",
  983. PRINT_TIMESTAMP_HEADER_WIDTH, "Max delay end");
  984. }
  985. printf("\n");
  986. print_separator(ret);
  987. return ret;
  988. }
  989. static void timehist_print_header(void)
  990. {
  991. /*
  992. * header row
  993. */
  994. printf(" %-*s %-*s %-*s %-*s %-*s %-*s\n",
  995. PRINT_TIMESTAMP_WIDTH, "Runtime start",
  996. PRINT_TIMESTAMP_WIDTH, "Runtime end",
  997. PRINT_TIMEHIST_CPU_WIDTH, "Cpu",
  998. PRINT_KWORK_NAME_WIDTH, "Kwork name",
  999. PRINT_RUNTIME_WIDTH, "Runtime",
  1000. PRINT_RUNTIME_WIDTH, "Delaytime");
  1001. /*
  1002. * units row
  1003. */
  1004. printf(" %-*s %-*s %-*s %-*s %-*s %-*s\n",
  1005. PRINT_TIMESTAMP_WIDTH, "",
  1006. PRINT_TIMESTAMP_WIDTH, "",
  1007. PRINT_TIMEHIST_CPU_WIDTH, "",
  1008. PRINT_KWORK_NAME_WIDTH, "(TYPE)NAME:NUM",
  1009. PRINT_RUNTIME_WIDTH, "(msec)",
  1010. PRINT_RUNTIME_WIDTH, "(msec)");
  1011. /*
  1012. * separator
  1013. */
  1014. printf(" %.*s %.*s %.*s %.*s %.*s %.*s\n",
  1015. PRINT_TIMESTAMP_WIDTH, graph_dotted_line,
  1016. PRINT_TIMESTAMP_WIDTH, graph_dotted_line,
  1017. PRINT_TIMEHIST_CPU_WIDTH, graph_dotted_line,
  1018. PRINT_KWORK_NAME_WIDTH, graph_dotted_line,
  1019. PRINT_RUNTIME_WIDTH, graph_dotted_line,
  1020. PRINT_RUNTIME_WIDTH, graph_dotted_line);
  1021. }
  1022. static void print_summary(struct perf_kwork *kwork)
  1023. {
  1024. u64 time = kwork->timeend - kwork->timestart;
  1025. printf(" Total count : %9" PRIu64 "\n", kwork->all_count);
  1026. printf(" Total runtime (msec) : %9.3f (%.3f%% load average)\n",
  1027. (double)kwork->all_runtime / NSEC_PER_MSEC,
  1028. time == 0 ? 0 : (double)kwork->all_runtime / time);
  1029. printf(" Total time span (msec) : %9.3f\n",
  1030. (double)time / NSEC_PER_MSEC);
  1031. }
  1032. static unsigned long long nr_list_entry(struct list_head *head)
  1033. {
  1034. struct list_head *pos;
  1035. unsigned long long n = 0;
  1036. list_for_each(pos, head)
  1037. n++;
  1038. return n;
  1039. }
  1040. static void print_skipped_events(struct perf_kwork *kwork)
  1041. {
  1042. int i;
  1043. const char *const kwork_event_str[] = {
  1044. [KWORK_TRACE_RAISE] = "raise",
  1045. [KWORK_TRACE_ENTRY] = "entry",
  1046. [KWORK_TRACE_EXIT] = "exit",
  1047. };
  1048. if ((kwork->nr_skipped_events[KWORK_TRACE_MAX] != 0) &&
  1049. (kwork->nr_events != 0)) {
  1050. printf(" INFO: %.3f%% skipped events (%" PRIu64 " including ",
  1051. (double)kwork->nr_skipped_events[KWORK_TRACE_MAX] /
  1052. (double)kwork->nr_events * 100.0,
  1053. kwork->nr_skipped_events[KWORK_TRACE_MAX]);
  1054. for (i = 0; i < KWORK_TRACE_MAX; i++) {
  1055. printf("%" PRIu64 " %s%s",
  1056. kwork->nr_skipped_events[i],
  1057. kwork_event_str[i],
  1058. (i == KWORK_TRACE_MAX - 1) ? ")\n" : ", ");
  1059. }
  1060. }
  1061. if (verbose > 0)
  1062. printf(" INFO: use %lld atom pages\n",
  1063. nr_list_entry(&kwork->atom_page_list));
  1064. }
  1065. static void print_bad_events(struct perf_kwork *kwork)
  1066. {
  1067. if ((kwork->nr_lost_events != 0) && (kwork->nr_events != 0)) {
  1068. printf(" INFO: %.3f%% lost events (%ld out of %ld, in %ld chunks)\n",
  1069. (double)kwork->nr_lost_events /
  1070. (double)kwork->nr_events * 100.0,
  1071. kwork->nr_lost_events, kwork->nr_events,
  1072. kwork->nr_lost_chunks);
  1073. }
  1074. }
  1075. static void work_sort(struct perf_kwork *kwork, struct kwork_class *class)
  1076. {
  1077. struct rb_node *node;
  1078. struct kwork_work *data;
  1079. struct rb_root_cached *root = &class->work_root;
  1080. pr_debug("Sorting %s ...\n", class->name);
  1081. for (;;) {
  1082. node = rb_first_cached(root);
  1083. if (!node)
  1084. break;
  1085. rb_erase_cached(node, root);
  1086. data = rb_entry(node, struct kwork_work, node);
  1087. work_insert(&kwork->sorted_work_root,
  1088. data, &kwork->sort_list);
  1089. }
  1090. }
  1091. static void perf_kwork__sort(struct perf_kwork *kwork)
  1092. {
  1093. struct kwork_class *class;
  1094. list_for_each_entry(class, &kwork->class_list, list)
  1095. work_sort(kwork, class);
  1096. }
  1097. static int perf_kwork__check_config(struct perf_kwork *kwork,
  1098. struct perf_session *session)
  1099. {
  1100. int ret;
  1101. struct evsel *evsel;
  1102. struct kwork_class *class;
  1103. static struct trace_kwork_handler report_ops = {
  1104. .entry_event = report_entry_event,
  1105. .exit_event = report_exit_event,
  1106. };
  1107. static struct trace_kwork_handler latency_ops = {
  1108. .raise_event = latency_raise_event,
  1109. .entry_event = latency_entry_event,
  1110. };
  1111. static struct trace_kwork_handler timehist_ops = {
  1112. .raise_event = timehist_raise_event,
  1113. .entry_event = timehist_entry_event,
  1114. .exit_event = timehist_exit_event,
  1115. };
  1116. switch (kwork->report) {
  1117. case KWORK_REPORT_RUNTIME:
  1118. kwork->tp_handler = &report_ops;
  1119. break;
  1120. case KWORK_REPORT_LATENCY:
  1121. kwork->tp_handler = &latency_ops;
  1122. break;
  1123. case KWORK_REPORT_TIMEHIST:
  1124. kwork->tp_handler = &timehist_ops;
  1125. break;
  1126. default:
  1127. pr_debug("Invalid report type %d\n", kwork->report);
  1128. return -1;
  1129. }
  1130. list_for_each_entry(class, &kwork->class_list, list)
  1131. if ((class->class_init != NULL) &&
  1132. (class->class_init(class, session) != 0))
  1133. return -1;
  1134. if (kwork->cpu_list != NULL) {
  1135. ret = perf_session__cpu_bitmap(session,
  1136. kwork->cpu_list,
  1137. kwork->cpu_bitmap);
  1138. if (ret < 0) {
  1139. pr_err("Invalid cpu bitmap\n");
  1140. return -1;
  1141. }
  1142. }
  1143. if (kwork->time_str != NULL) {
  1144. ret = perf_time__parse_str(&kwork->ptime, kwork->time_str);
  1145. if (ret != 0) {
  1146. pr_err("Invalid time span\n");
  1147. return -1;
  1148. }
  1149. }
  1150. list_for_each_entry(evsel, &session->evlist->core.entries, core.node) {
  1151. if (kwork->show_callchain && !evsel__has_callchain(evsel)) {
  1152. pr_debug("Samples do not have callchains\n");
  1153. kwork->show_callchain = 0;
  1154. symbol_conf.use_callchain = 0;
  1155. }
  1156. }
  1157. return 0;
  1158. }
  1159. static int perf_kwork__read_events(struct perf_kwork *kwork)
  1160. {
  1161. int ret = -1;
  1162. struct perf_session *session = NULL;
  1163. struct perf_data data = {
  1164. .path = input_name,
  1165. .mode = PERF_DATA_MODE_READ,
  1166. .force = kwork->force,
  1167. };
  1168. session = perf_session__new(&data, &kwork->tool);
  1169. if (IS_ERR(session)) {
  1170. pr_debug("Error creating perf session\n");
  1171. return PTR_ERR(session);
  1172. }
  1173. symbol__init(&session->header.env);
  1174. if (perf_kwork__check_config(kwork, session) != 0)
  1175. goto out_delete;
  1176. if (session->tevent.pevent &&
  1177. tep_set_function_resolver(session->tevent.pevent,
  1178. machine__resolve_kernel_addr,
  1179. &session->machines.host) < 0) {
  1180. pr_err("Failed to set libtraceevent function resolver\n");
  1181. goto out_delete;
  1182. }
  1183. if (kwork->report == KWORK_REPORT_TIMEHIST)
  1184. timehist_print_header();
  1185. ret = perf_session__process_events(session);
  1186. if (ret) {
  1187. pr_debug("Failed to process events, error %d\n", ret);
  1188. goto out_delete;
  1189. }
  1190. kwork->nr_events = session->evlist->stats.nr_events[0];
  1191. kwork->nr_lost_events = session->evlist->stats.total_lost;
  1192. kwork->nr_lost_chunks = session->evlist->stats.nr_events[PERF_RECORD_LOST];
  1193. out_delete:
  1194. perf_session__delete(session);
  1195. return ret;
  1196. }
  1197. static void process_skipped_events(struct perf_kwork *kwork,
  1198. struct kwork_work *work)
  1199. {
  1200. int i;
  1201. unsigned long long count;
  1202. for (i = 0; i < KWORK_TRACE_MAX; i++) {
  1203. count = nr_list_entry(&work->atom_list[i]);
  1204. kwork->nr_skipped_events[i] += count;
  1205. kwork->nr_skipped_events[KWORK_TRACE_MAX] += count;
  1206. }
  1207. }
  1208. struct kwork_work *perf_kwork_add_work(struct perf_kwork *kwork,
  1209. struct kwork_class *class,
  1210. struct kwork_work *key)
  1211. {
  1212. struct kwork_work *work = NULL;
  1213. work = work_new(key);
  1214. if (work == NULL)
  1215. return NULL;
  1216. work_insert(&class->work_root, work, &kwork->cmp_id);
  1217. return work;
  1218. }
  1219. static void sig_handler(int sig)
  1220. {
  1221. /*
  1222. * Simply capture termination signal so that
  1223. * the program can continue after pause returns
  1224. */
  1225. pr_debug("Captuer signal %d\n", sig);
  1226. }
  1227. static int perf_kwork__report_bpf(struct perf_kwork *kwork)
  1228. {
  1229. int ret;
  1230. signal(SIGINT, sig_handler);
  1231. signal(SIGTERM, sig_handler);
  1232. ret = perf_kwork__trace_prepare_bpf(kwork);
  1233. if (ret)
  1234. return -1;
  1235. printf("Starting trace, Hit <Ctrl+C> to stop and report\n");
  1236. perf_kwork__trace_start();
  1237. /*
  1238. * a simple pause, wait here for stop signal
  1239. */
  1240. pause();
  1241. perf_kwork__trace_finish();
  1242. perf_kwork__report_read_bpf(kwork);
  1243. perf_kwork__report_cleanup_bpf();
  1244. return 0;
  1245. }
  1246. static int perf_kwork__report(struct perf_kwork *kwork)
  1247. {
  1248. int ret;
  1249. struct rb_node *next;
  1250. struct kwork_work *work;
  1251. if (kwork->use_bpf)
  1252. ret = perf_kwork__report_bpf(kwork);
  1253. else
  1254. ret = perf_kwork__read_events(kwork);
  1255. if (ret != 0)
  1256. return -1;
  1257. perf_kwork__sort(kwork);
  1258. setup_pager();
  1259. ret = report_print_header(kwork);
  1260. next = rb_first_cached(&kwork->sorted_work_root);
  1261. while (next) {
  1262. work = rb_entry(next, struct kwork_work, node);
  1263. process_skipped_events(kwork, work);
  1264. if (work->nr_atoms != 0) {
  1265. report_print_work(kwork, work);
  1266. if (kwork->summary) {
  1267. kwork->all_runtime += work->total_runtime;
  1268. kwork->all_count += work->nr_atoms;
  1269. }
  1270. }
  1271. next = rb_next(next);
  1272. }
  1273. print_separator(ret);
  1274. if (kwork->summary) {
  1275. print_summary(kwork);
  1276. print_separator(ret);
  1277. }
  1278. print_bad_events(kwork);
  1279. print_skipped_events(kwork);
  1280. printf("\n");
  1281. return 0;
  1282. }
  1283. typedef int (*tracepoint_handler)(struct perf_tool *tool,
  1284. struct evsel *evsel,
  1285. struct perf_sample *sample,
  1286. struct machine *machine);
  1287. static int perf_kwork__process_tracepoint_sample(struct perf_tool *tool,
  1288. union perf_event *event __maybe_unused,
  1289. struct perf_sample *sample,
  1290. struct evsel *evsel,
  1291. struct machine *machine)
  1292. {
  1293. int err = 0;
  1294. if (evsel->handler != NULL) {
  1295. tracepoint_handler f = evsel->handler;
  1296. err = f(tool, evsel, sample, machine);
  1297. }
  1298. return err;
  1299. }
  1300. static int perf_kwork__timehist(struct perf_kwork *kwork)
  1301. {
  1302. /*
  1303. * event handlers for timehist option
  1304. */
  1305. kwork->tool.comm = perf_event__process_comm;
  1306. kwork->tool.exit = perf_event__process_exit;
  1307. kwork->tool.fork = perf_event__process_fork;
  1308. kwork->tool.attr = perf_event__process_attr;
  1309. kwork->tool.tracing_data = perf_event__process_tracing_data;
  1310. kwork->tool.build_id = perf_event__process_build_id;
  1311. kwork->tool.ordered_events = true;
  1312. kwork->tool.ordering_requires_timestamps = true;
  1313. symbol_conf.use_callchain = kwork->show_callchain;
  1314. if (symbol__validate_sym_arguments()) {
  1315. pr_err("Failed to validate sym arguments\n");
  1316. return -1;
  1317. }
  1318. setup_pager();
  1319. return perf_kwork__read_events(kwork);
  1320. }
  1321. static void setup_event_list(struct perf_kwork *kwork,
  1322. const struct option *options,
  1323. const char * const usage_msg[])
  1324. {
  1325. int i;
  1326. struct kwork_class *class;
  1327. char *tmp, *tok, *str;
  1328. if (kwork->event_list_str == NULL)
  1329. goto null_event_list_str;
  1330. str = strdup(kwork->event_list_str);
  1331. for (tok = strtok_r(str, ", ", &tmp);
  1332. tok; tok = strtok_r(NULL, ", ", &tmp)) {
  1333. for (i = 0; i < KWORK_CLASS_MAX; i++) {
  1334. class = kwork_class_supported_list[i];
  1335. if (strcmp(tok, class->name) == 0) {
  1336. list_add_tail(&class->list, &kwork->class_list);
  1337. break;
  1338. }
  1339. }
  1340. if (i == KWORK_CLASS_MAX) {
  1341. usage_with_options_msg(usage_msg, options,
  1342. "Unknown --event key: `%s'", tok);
  1343. }
  1344. }
  1345. free(str);
  1346. null_event_list_str:
  1347. /*
  1348. * config all kwork events if not specified
  1349. */
  1350. if (list_empty(&kwork->class_list)) {
  1351. for (i = 0; i < KWORK_CLASS_MAX; i++) {
  1352. list_add_tail(&kwork_class_supported_list[i]->list,
  1353. &kwork->class_list);
  1354. }
  1355. }
  1356. pr_debug("Config event list:");
  1357. list_for_each_entry(class, &kwork->class_list, list)
  1358. pr_debug(" %s", class->name);
  1359. pr_debug("\n");
  1360. }
  1361. static int perf_kwork__record(struct perf_kwork *kwork,
  1362. int argc, const char **argv)
  1363. {
  1364. const char **rec_argv;
  1365. unsigned int rec_argc, i, j;
  1366. struct kwork_class *class;
  1367. const char *const record_args[] = {
  1368. "record",
  1369. "-a",
  1370. "-R",
  1371. "-m", "1024",
  1372. "-c", "1",
  1373. };
  1374. rec_argc = ARRAY_SIZE(record_args) + argc - 1;
  1375. list_for_each_entry(class, &kwork->class_list, list)
  1376. rec_argc += 2 * class->nr_tracepoints;
  1377. rec_argv = calloc(rec_argc + 1, sizeof(char *));
  1378. if (rec_argv == NULL)
  1379. return -ENOMEM;
  1380. for (i = 0; i < ARRAY_SIZE(record_args); i++)
  1381. rec_argv[i] = strdup(record_args[i]);
  1382. list_for_each_entry(class, &kwork->class_list, list) {
  1383. for (j = 0; j < class->nr_tracepoints; j++) {
  1384. rec_argv[i++] = strdup("-e");
  1385. rec_argv[i++] = strdup(class->tp_handlers[j].name);
  1386. }
  1387. }
  1388. for (j = 1; j < (unsigned int)argc; j++, i++)
  1389. rec_argv[i] = argv[j];
  1390. BUG_ON(i != rec_argc);
  1391. pr_debug("record comm: ");
  1392. for (j = 0; j < rec_argc; j++)
  1393. pr_debug("%s ", rec_argv[j]);
  1394. pr_debug("\n");
  1395. return cmd_record(i, rec_argv);
  1396. }
  1397. int cmd_kwork(int argc, const char **argv)
  1398. {
  1399. static struct perf_kwork kwork = {
  1400. .class_list = LIST_HEAD_INIT(kwork.class_list),
  1401. .tool = {
  1402. .mmap = perf_event__process_mmap,
  1403. .mmap2 = perf_event__process_mmap2,
  1404. .sample = perf_kwork__process_tracepoint_sample,
  1405. .ordered_events = true,
  1406. },
  1407. .atom_page_list = LIST_HEAD_INIT(kwork.atom_page_list),
  1408. .sort_list = LIST_HEAD_INIT(kwork.sort_list),
  1409. .cmp_id = LIST_HEAD_INIT(kwork.cmp_id),
  1410. .sorted_work_root = RB_ROOT_CACHED,
  1411. .tp_handler = NULL,
  1412. .profile_name = NULL,
  1413. .cpu_list = NULL,
  1414. .time_str = NULL,
  1415. .force = false,
  1416. .event_list_str = NULL,
  1417. .summary = false,
  1418. .sort_order = NULL,
  1419. .show_callchain = false,
  1420. .max_stack = 5,
  1421. .timestart = 0,
  1422. .timeend = 0,
  1423. .nr_events = 0,
  1424. .nr_lost_chunks = 0,
  1425. .nr_lost_events = 0,
  1426. .all_runtime = 0,
  1427. .all_count = 0,
  1428. .nr_skipped_events = { 0 },
  1429. };
  1430. static const char default_report_sort_order[] = "runtime, max, count";
  1431. static const char default_latency_sort_order[] = "avg, max, count";
  1432. const struct option kwork_options[] = {
  1433. OPT_INCR('v', "verbose", &verbose,
  1434. "be more verbose (show symbol address, etc)"),
  1435. OPT_BOOLEAN('D', "dump-raw-trace", &dump_trace,
  1436. "dump raw trace in ASCII"),
  1437. OPT_STRING('k', "kwork", &kwork.event_list_str, "kwork",
  1438. "list of kwork to profile (irq, softirq, workqueue, etc)"),
  1439. OPT_BOOLEAN('f', "force", &kwork.force, "don't complain, do it"),
  1440. OPT_END()
  1441. };
  1442. const struct option report_options[] = {
  1443. OPT_STRING('s', "sort", &kwork.sort_order, "key[,key2...]",
  1444. "sort by key(s): runtime, max, count"),
  1445. OPT_STRING('C', "cpu", &kwork.cpu_list, "cpu",
  1446. "list of cpus to profile"),
  1447. OPT_STRING('n', "name", &kwork.profile_name, "name",
  1448. "event name to profile"),
  1449. OPT_STRING(0, "time", &kwork.time_str, "str",
  1450. "Time span for analysis (start,stop)"),
  1451. OPT_STRING('i', "input", &input_name, "file",
  1452. "input file name"),
  1453. OPT_BOOLEAN('S', "with-summary", &kwork.summary,
  1454. "Show summary with statistics"),
  1455. #ifdef HAVE_BPF_SKEL
  1456. OPT_BOOLEAN('b', "use-bpf", &kwork.use_bpf,
  1457. "Use BPF to measure kwork runtime"),
  1458. #endif
  1459. OPT_PARENT(kwork_options)
  1460. };
  1461. const struct option latency_options[] = {
  1462. OPT_STRING('s', "sort", &kwork.sort_order, "key[,key2...]",
  1463. "sort by key(s): avg, max, count"),
  1464. OPT_STRING('C', "cpu", &kwork.cpu_list, "cpu",
  1465. "list of cpus to profile"),
  1466. OPT_STRING('n', "name", &kwork.profile_name, "name",
  1467. "event name to profile"),
  1468. OPT_STRING(0, "time", &kwork.time_str, "str",
  1469. "Time span for analysis (start,stop)"),
  1470. OPT_STRING('i', "input", &input_name, "file",
  1471. "input file name"),
  1472. #ifdef HAVE_BPF_SKEL
  1473. OPT_BOOLEAN('b', "use-bpf", &kwork.use_bpf,
  1474. "Use BPF to measure kwork latency"),
  1475. #endif
  1476. OPT_PARENT(kwork_options)
  1477. };
  1478. const struct option timehist_options[] = {
  1479. OPT_STRING('k', "vmlinux", &symbol_conf.vmlinux_name,
  1480. "file", "vmlinux pathname"),
  1481. OPT_STRING(0, "kallsyms", &symbol_conf.kallsyms_name,
  1482. "file", "kallsyms pathname"),
  1483. OPT_BOOLEAN('g', "call-graph", &kwork.show_callchain,
  1484. "Display call chains if present"),
  1485. OPT_UINTEGER(0, "max-stack", &kwork.max_stack,
  1486. "Maximum number of functions to display backtrace."),
  1487. OPT_STRING(0, "symfs", &symbol_conf.symfs, "directory",
  1488. "Look for files with symbols relative to this directory"),
  1489. OPT_STRING(0, "time", &kwork.time_str, "str",
  1490. "Time span for analysis (start,stop)"),
  1491. OPT_STRING('C', "cpu", &kwork.cpu_list, "cpu",
  1492. "list of cpus to profile"),
  1493. OPT_STRING('n', "name", &kwork.profile_name, "name",
  1494. "event name to profile"),
  1495. OPT_STRING('i', "input", &input_name, "file",
  1496. "input file name"),
  1497. OPT_PARENT(kwork_options)
  1498. };
  1499. const char *kwork_usage[] = {
  1500. NULL,
  1501. NULL
  1502. };
  1503. const char * const report_usage[] = {
  1504. "perf kwork report [<options>]",
  1505. NULL
  1506. };
  1507. const char * const latency_usage[] = {
  1508. "perf kwork latency [<options>]",
  1509. NULL
  1510. };
  1511. const char * const timehist_usage[] = {
  1512. "perf kwork timehist [<options>]",
  1513. NULL
  1514. };
  1515. const char *const kwork_subcommands[] = {
  1516. "record", "report", "latency", "timehist", NULL
  1517. };
  1518. argc = parse_options_subcommand(argc, argv, kwork_options,
  1519. kwork_subcommands, kwork_usage,
  1520. PARSE_OPT_STOP_AT_NON_OPTION);
  1521. if (!argc)
  1522. usage_with_options(kwork_usage, kwork_options);
  1523. setup_event_list(&kwork, kwork_options, kwork_usage);
  1524. sort_dimension__add(&kwork, "id", &kwork.cmp_id);
  1525. if (strlen(argv[0]) > 2 && strstarts("record", argv[0]))
  1526. return perf_kwork__record(&kwork, argc, argv);
  1527. else if (strlen(argv[0]) > 2 && strstarts("report", argv[0])) {
  1528. kwork.sort_order = default_report_sort_order;
  1529. if (argc > 1) {
  1530. argc = parse_options(argc, argv, report_options, report_usage, 0);
  1531. if (argc)
  1532. usage_with_options(report_usage, report_options);
  1533. }
  1534. kwork.report = KWORK_REPORT_RUNTIME;
  1535. setup_sorting(&kwork, report_options, report_usage);
  1536. return perf_kwork__report(&kwork);
  1537. } else if (strlen(argv[0]) > 2 && strstarts("latency", argv[0])) {
  1538. kwork.sort_order = default_latency_sort_order;
  1539. if (argc > 1) {
  1540. argc = parse_options(argc, argv, latency_options, latency_usage, 0);
  1541. if (argc)
  1542. usage_with_options(latency_usage, latency_options);
  1543. }
  1544. kwork.report = KWORK_REPORT_LATENCY;
  1545. setup_sorting(&kwork, latency_options, latency_usage);
  1546. return perf_kwork__report(&kwork);
  1547. } else if (strlen(argv[0]) > 2 && strstarts("timehist", argv[0])) {
  1548. if (argc > 1) {
  1549. argc = parse_options(argc, argv, timehist_options, timehist_usage, 0);
  1550. if (argc)
  1551. usage_with_options(timehist_usage, timehist_options);
  1552. }
  1553. kwork.report = KWORK_REPORT_TIMEHIST;
  1554. return perf_kwork__timehist(&kwork);
  1555. } else
  1556. usage_with_options(kwork_usage, kwork_options);
  1557. return 0;
  1558. }