xhci-trace.h 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631
  1. /* SPDX-License-Identifier: GPL-2.0 */
  2. /*
  3. * xHCI host controller driver
  4. *
  5. * Copyright (C) 2013 Xenia Ragiadakou
  6. *
  7. * Author: Xenia Ragiadakou
  8. * Email : [email protected]
  9. */
  10. #undef TRACE_SYSTEM
  11. #define TRACE_SYSTEM xhci-hcd
  12. /*
  13. * The TRACE_SYSTEM_VAR defaults to TRACE_SYSTEM, but must be a
  14. * legitimate C variable. It is not exported to user space.
  15. */
  16. #undef TRACE_SYSTEM_VAR
  17. #define TRACE_SYSTEM_VAR xhci_hcd
  18. #if !defined(__XHCI_TRACE_H) || defined(TRACE_HEADER_MULTI_READ)
  19. #define __XHCI_TRACE_H
  20. #include <linux/tracepoint.h>
  21. #include "xhci.h"
  22. #include "xhci-dbgcap.h"
  23. DECLARE_EVENT_CLASS(xhci_log_msg,
  24. TP_PROTO(struct va_format *vaf),
  25. TP_ARGS(vaf),
  26. TP_STRUCT__entry(__vstring(msg, vaf->fmt, vaf->va)),
  27. TP_fast_assign(
  28. __assign_vstr(msg, vaf->fmt, vaf->va);
  29. ),
  30. TP_printk("%s", __get_str(msg))
  31. );
  32. DEFINE_EVENT(xhci_log_msg, xhci_dbg_address,
  33. TP_PROTO(struct va_format *vaf),
  34. TP_ARGS(vaf)
  35. );
  36. DEFINE_EVENT(xhci_log_msg, xhci_dbg_context_change,
  37. TP_PROTO(struct va_format *vaf),
  38. TP_ARGS(vaf)
  39. );
  40. DEFINE_EVENT(xhci_log_msg, xhci_dbg_quirks,
  41. TP_PROTO(struct va_format *vaf),
  42. TP_ARGS(vaf)
  43. );
  44. DEFINE_EVENT(xhci_log_msg, xhci_dbg_reset_ep,
  45. TP_PROTO(struct va_format *vaf),
  46. TP_ARGS(vaf)
  47. );
  48. DEFINE_EVENT(xhci_log_msg, xhci_dbg_cancel_urb,
  49. TP_PROTO(struct va_format *vaf),
  50. TP_ARGS(vaf)
  51. );
  52. DEFINE_EVENT(xhci_log_msg, xhci_dbg_init,
  53. TP_PROTO(struct va_format *vaf),
  54. TP_ARGS(vaf)
  55. );
  56. DEFINE_EVENT(xhci_log_msg, xhci_dbg_ring_expansion,
  57. TP_PROTO(struct va_format *vaf),
  58. TP_ARGS(vaf)
  59. );
  60. DECLARE_EVENT_CLASS(xhci_log_ctx,
  61. TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx,
  62. unsigned int ep_num),
  63. TP_ARGS(xhci, ctx, ep_num),
  64. TP_STRUCT__entry(
  65. __field(int, ctx_64)
  66. __field(unsigned, ctx_type)
  67. __field(dma_addr_t, ctx_dma)
  68. __field(u8 *, ctx_va)
  69. __field(unsigned, ctx_ep_num)
  70. __dynamic_array(u32, ctx_data,
  71. ((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 8) *
  72. ((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1))
  73. ),
  74. TP_fast_assign(
  75. __entry->ctx_64 = HCC_64BYTE_CONTEXT(xhci->hcc_params);
  76. __entry->ctx_type = ctx->type;
  77. __entry->ctx_dma = ctx->dma;
  78. __entry->ctx_va = ctx->bytes;
  79. __entry->ctx_ep_num = ep_num;
  80. memcpy(__get_dynamic_array(ctx_data), ctx->bytes,
  81. ((HCC_64BYTE_CONTEXT(xhci->hcc_params) + 1) * 32) *
  82. ((ctx->type == XHCI_CTX_TYPE_INPUT) + ep_num + 1));
  83. ),
  84. TP_printk("ctx_64=%d, ctx_type=%u, ctx_dma=@%llx, ctx_va=@%p",
  85. __entry->ctx_64, __entry->ctx_type,
  86. (unsigned long long) __entry->ctx_dma, __entry->ctx_va
  87. )
  88. );
  89. DEFINE_EVENT(xhci_log_ctx, xhci_address_ctx,
  90. TP_PROTO(struct xhci_hcd *xhci, struct xhci_container_ctx *ctx,
  91. unsigned int ep_num),
  92. TP_ARGS(xhci, ctx, ep_num)
  93. );
  94. DECLARE_EVENT_CLASS(xhci_log_trb,
  95. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  96. TP_ARGS(ring, trb),
  97. TP_STRUCT__entry(
  98. __field(u32, type)
  99. __field(u32, field0)
  100. __field(u32, field1)
  101. __field(u32, field2)
  102. __field(u32, field3)
  103. __dynamic_array(char, str, XHCI_MSG_MAX)
  104. ),
  105. TP_fast_assign(
  106. __entry->type = ring->type;
  107. __entry->field0 = le32_to_cpu(trb->field[0]);
  108. __entry->field1 = le32_to_cpu(trb->field[1]);
  109. __entry->field2 = le32_to_cpu(trb->field[2]);
  110. __entry->field3 = le32_to_cpu(trb->field[3]);
  111. ),
  112. TP_printk("%s: %s", xhci_ring_type_string(__entry->type),
  113. xhci_decode_trb(__get_str(str), XHCI_MSG_MAX, __entry->field0, __entry->field1,
  114. __entry->field2, __entry->field3)
  115. )
  116. );
  117. DEFINE_EVENT(xhci_log_trb, xhci_handle_event,
  118. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  119. TP_ARGS(ring, trb)
  120. );
  121. DEFINE_EVENT(xhci_log_trb, xhci_handle_command,
  122. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  123. TP_ARGS(ring, trb)
  124. );
  125. DEFINE_EVENT(xhci_log_trb, xhci_handle_transfer,
  126. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  127. TP_ARGS(ring, trb)
  128. );
  129. DEFINE_EVENT(xhci_log_trb, xhci_queue_trb,
  130. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  131. TP_ARGS(ring, trb)
  132. );
  133. DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_event,
  134. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  135. TP_ARGS(ring, trb)
  136. );
  137. DEFINE_EVENT(xhci_log_trb, xhci_dbc_handle_transfer,
  138. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  139. TP_ARGS(ring, trb)
  140. );
  141. DEFINE_EVENT(xhci_log_trb, xhci_dbc_gadget_ep_queue,
  142. TP_PROTO(struct xhci_ring *ring, struct xhci_generic_trb *trb),
  143. TP_ARGS(ring, trb)
  144. );
  145. DECLARE_EVENT_CLASS(xhci_log_free_virt_dev,
  146. TP_PROTO(struct xhci_virt_device *vdev),
  147. TP_ARGS(vdev),
  148. TP_STRUCT__entry(
  149. __field(void *, vdev)
  150. __field(unsigned long long, out_ctx)
  151. __field(unsigned long long, in_ctx)
  152. __field(u8, fake_port)
  153. __field(u8, real_port)
  154. __field(u16, current_mel)
  155. ),
  156. TP_fast_assign(
  157. __entry->vdev = vdev;
  158. __entry->in_ctx = (unsigned long long) vdev->in_ctx->dma;
  159. __entry->out_ctx = (unsigned long long) vdev->out_ctx->dma;
  160. __entry->fake_port = (u8) vdev->fake_port;
  161. __entry->real_port = (u8) vdev->real_port;
  162. __entry->current_mel = (u16) vdev->current_mel;
  163. ),
  164. TP_printk("vdev %p ctx %llx | %llx fake_port %d real_port %d current_mel %d",
  165. __entry->vdev, __entry->in_ctx, __entry->out_ctx,
  166. __entry->fake_port, __entry->real_port, __entry->current_mel
  167. )
  168. );
  169. DEFINE_EVENT(xhci_log_free_virt_dev, xhci_free_virt_device,
  170. TP_PROTO(struct xhci_virt_device *vdev),
  171. TP_ARGS(vdev)
  172. );
  173. DECLARE_EVENT_CLASS(xhci_log_virt_dev,
  174. TP_PROTO(struct xhci_virt_device *vdev),
  175. TP_ARGS(vdev),
  176. TP_STRUCT__entry(
  177. __field(void *, vdev)
  178. __field(unsigned long long, out_ctx)
  179. __field(unsigned long long, in_ctx)
  180. __field(int, devnum)
  181. __field(int, state)
  182. __field(int, speed)
  183. __field(u8, portnum)
  184. __field(u8, level)
  185. __field(int, slot_id)
  186. ),
  187. TP_fast_assign(
  188. __entry->vdev = vdev;
  189. __entry->in_ctx = (unsigned long long) vdev->in_ctx->dma;
  190. __entry->out_ctx = (unsigned long long) vdev->out_ctx->dma;
  191. __entry->devnum = vdev->udev->devnum;
  192. __entry->state = vdev->udev->state;
  193. __entry->speed = vdev->udev->speed;
  194. __entry->portnum = vdev->udev->portnum;
  195. __entry->level = vdev->udev->level;
  196. __entry->slot_id = vdev->udev->slot_id;
  197. ),
  198. TP_printk("vdev %p ctx %llx | %llx num %d state %d speed %d port %d level %d slot %d",
  199. __entry->vdev, __entry->in_ctx, __entry->out_ctx,
  200. __entry->devnum, __entry->state, __entry->speed,
  201. __entry->portnum, __entry->level, __entry->slot_id
  202. )
  203. );
  204. DEFINE_EVENT(xhci_log_virt_dev, xhci_alloc_virt_device,
  205. TP_PROTO(struct xhci_virt_device *vdev),
  206. TP_ARGS(vdev)
  207. );
  208. DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_device,
  209. TP_PROTO(struct xhci_virt_device *vdev),
  210. TP_ARGS(vdev)
  211. );
  212. DEFINE_EVENT(xhci_log_virt_dev, xhci_setup_addressable_virt_device,
  213. TP_PROTO(struct xhci_virt_device *vdev),
  214. TP_ARGS(vdev)
  215. );
  216. DEFINE_EVENT(xhci_log_virt_dev, xhci_stop_device,
  217. TP_PROTO(struct xhci_virt_device *vdev),
  218. TP_ARGS(vdev)
  219. );
  220. DECLARE_EVENT_CLASS(xhci_log_urb,
  221. TP_PROTO(struct urb *urb),
  222. TP_ARGS(urb),
  223. TP_STRUCT__entry(
  224. __field(void *, urb)
  225. __field(unsigned int, pipe)
  226. __field(unsigned int, stream)
  227. __field(int, status)
  228. __field(unsigned int, flags)
  229. __field(int, num_mapped_sgs)
  230. __field(int, num_sgs)
  231. __field(int, length)
  232. __field(int, actual)
  233. __field(int, epnum)
  234. __field(int, dir_in)
  235. __field(int, type)
  236. __field(int, slot_id)
  237. ),
  238. TP_fast_assign(
  239. __entry->urb = urb;
  240. __entry->pipe = urb->pipe;
  241. __entry->stream = urb->stream_id;
  242. __entry->status = urb->status;
  243. __entry->flags = urb->transfer_flags;
  244. __entry->num_mapped_sgs = urb->num_mapped_sgs;
  245. __entry->num_sgs = urb->num_sgs;
  246. __entry->length = urb->transfer_buffer_length;
  247. __entry->actual = urb->actual_length;
  248. __entry->epnum = usb_endpoint_num(&urb->ep->desc);
  249. __entry->dir_in = usb_endpoint_dir_in(&urb->ep->desc);
  250. __entry->type = usb_endpoint_type(&urb->ep->desc);
  251. __entry->slot_id = urb->dev->slot_id;
  252. ),
  253. TP_printk("ep%d%s-%s: urb %p pipe %u slot %d length %d/%d sgs %d/%d stream %d flags %08x",
  254. __entry->epnum, __entry->dir_in ? "in" : "out",
  255. __print_symbolic(__entry->type,
  256. { USB_ENDPOINT_XFER_INT, "intr" },
  257. { USB_ENDPOINT_XFER_CONTROL, "control" },
  258. { USB_ENDPOINT_XFER_BULK, "bulk" },
  259. { USB_ENDPOINT_XFER_ISOC, "isoc" }),
  260. __entry->urb, __entry->pipe, __entry->slot_id,
  261. __entry->actual, __entry->length, __entry->num_mapped_sgs,
  262. __entry->num_sgs, __entry->stream, __entry->flags
  263. )
  264. );
  265. DEFINE_EVENT(xhci_log_urb, xhci_urb_enqueue,
  266. TP_PROTO(struct urb *urb),
  267. TP_ARGS(urb)
  268. );
  269. DEFINE_EVENT(xhci_log_urb, xhci_urb_giveback,
  270. TP_PROTO(struct urb *urb),
  271. TP_ARGS(urb)
  272. );
  273. DEFINE_EVENT(xhci_log_urb, xhci_urb_dequeue,
  274. TP_PROTO(struct urb *urb),
  275. TP_ARGS(urb)
  276. );
  277. DECLARE_EVENT_CLASS(xhci_log_ep_ctx,
  278. TP_PROTO(struct xhci_ep_ctx *ctx),
  279. TP_ARGS(ctx),
  280. TP_STRUCT__entry(
  281. __field(u32, info)
  282. __field(u32, info2)
  283. __field(u64, deq)
  284. __field(u32, tx_info)
  285. __dynamic_array(char, str, XHCI_MSG_MAX)
  286. ),
  287. TP_fast_assign(
  288. __entry->info = le32_to_cpu(ctx->ep_info);
  289. __entry->info2 = le32_to_cpu(ctx->ep_info2);
  290. __entry->deq = le64_to_cpu(ctx->deq);
  291. __entry->tx_info = le32_to_cpu(ctx->tx_info);
  292. ),
  293. TP_printk("%s", xhci_decode_ep_context(__get_str(str),
  294. __entry->info, __entry->info2, __entry->deq, __entry->tx_info)
  295. )
  296. );
  297. DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_stop_ep,
  298. TP_PROTO(struct xhci_ep_ctx *ctx),
  299. TP_ARGS(ctx)
  300. );
  301. DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_set_deq_ep,
  302. TP_PROTO(struct xhci_ep_ctx *ctx),
  303. TP_ARGS(ctx)
  304. );
  305. DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_reset_ep,
  306. TP_PROTO(struct xhci_ep_ctx *ctx),
  307. TP_ARGS(ctx)
  308. );
  309. DEFINE_EVENT(xhci_log_ep_ctx, xhci_handle_cmd_config_ep,
  310. TP_PROTO(struct xhci_ep_ctx *ctx),
  311. TP_ARGS(ctx)
  312. );
  313. DEFINE_EVENT(xhci_log_ep_ctx, xhci_add_endpoint,
  314. TP_PROTO(struct xhci_ep_ctx *ctx),
  315. TP_ARGS(ctx)
  316. );
  317. DECLARE_EVENT_CLASS(xhci_log_slot_ctx,
  318. TP_PROTO(struct xhci_slot_ctx *ctx),
  319. TP_ARGS(ctx),
  320. TP_STRUCT__entry(
  321. __field(u32, info)
  322. __field(u32, info2)
  323. __field(u32, tt_info)
  324. __field(u32, state)
  325. __dynamic_array(char, str, XHCI_MSG_MAX)
  326. ),
  327. TP_fast_assign(
  328. __entry->info = le32_to_cpu(ctx->dev_info);
  329. __entry->info2 = le32_to_cpu(ctx->dev_info2);
  330. __entry->tt_info = le64_to_cpu(ctx->tt_info);
  331. __entry->state = le32_to_cpu(ctx->dev_state);
  332. ),
  333. TP_printk("%s", xhci_decode_slot_context(__get_str(str),
  334. __entry->info, __entry->info2,
  335. __entry->tt_info, __entry->state)
  336. )
  337. );
  338. DEFINE_EVENT(xhci_log_slot_ctx, xhci_alloc_dev,
  339. TP_PROTO(struct xhci_slot_ctx *ctx),
  340. TP_ARGS(ctx)
  341. );
  342. DEFINE_EVENT(xhci_log_slot_ctx, xhci_free_dev,
  343. TP_PROTO(struct xhci_slot_ctx *ctx),
  344. TP_ARGS(ctx)
  345. );
  346. DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_disable_slot,
  347. TP_PROTO(struct xhci_slot_ctx *ctx),
  348. TP_ARGS(ctx)
  349. );
  350. DEFINE_EVENT(xhci_log_slot_ctx, xhci_discover_or_reset_device,
  351. TP_PROTO(struct xhci_slot_ctx *ctx),
  352. TP_ARGS(ctx)
  353. );
  354. DEFINE_EVENT(xhci_log_slot_ctx, xhci_setup_device_slot,
  355. TP_PROTO(struct xhci_slot_ctx *ctx),
  356. TP_ARGS(ctx)
  357. );
  358. DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_addr_dev,
  359. TP_PROTO(struct xhci_slot_ctx *ctx),
  360. TP_ARGS(ctx)
  361. );
  362. DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_reset_dev,
  363. TP_PROTO(struct xhci_slot_ctx *ctx),
  364. TP_ARGS(ctx)
  365. );
  366. DEFINE_EVENT(xhci_log_slot_ctx, xhci_handle_cmd_set_deq,
  367. TP_PROTO(struct xhci_slot_ctx *ctx),
  368. TP_ARGS(ctx)
  369. );
  370. DEFINE_EVENT(xhci_log_slot_ctx, xhci_configure_endpoint,
  371. TP_PROTO(struct xhci_slot_ctx *ctx),
  372. TP_ARGS(ctx)
  373. );
  374. DECLARE_EVENT_CLASS(xhci_log_ctrl_ctx,
  375. TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
  376. TP_ARGS(ctrl_ctx),
  377. TP_STRUCT__entry(
  378. __field(u32, drop)
  379. __field(u32, add)
  380. __dynamic_array(char, str, XHCI_MSG_MAX)
  381. ),
  382. TP_fast_assign(
  383. __entry->drop = le32_to_cpu(ctrl_ctx->drop_flags);
  384. __entry->add = le32_to_cpu(ctrl_ctx->add_flags);
  385. ),
  386. TP_printk("%s", xhci_decode_ctrl_ctx(__get_str(str), __entry->drop, __entry->add)
  387. )
  388. );
  389. DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_address_ctrl_ctx,
  390. TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
  391. TP_ARGS(ctrl_ctx)
  392. );
  393. DEFINE_EVENT(xhci_log_ctrl_ctx, xhci_configure_endpoint_ctrl_ctx,
  394. TP_PROTO(struct xhci_input_control_ctx *ctrl_ctx),
  395. TP_ARGS(ctrl_ctx)
  396. );
  397. DECLARE_EVENT_CLASS(xhci_log_ring,
  398. TP_PROTO(struct xhci_ring *ring),
  399. TP_ARGS(ring),
  400. TP_STRUCT__entry(
  401. __field(u32, type)
  402. __field(void *, ring)
  403. __field(dma_addr_t, enq)
  404. __field(dma_addr_t, deq)
  405. __field(dma_addr_t, enq_seg)
  406. __field(dma_addr_t, deq_seg)
  407. __field(unsigned int, num_segs)
  408. __field(unsigned int, stream_id)
  409. __field(unsigned int, cycle_state)
  410. __field(unsigned int, num_trbs_free)
  411. __field(unsigned int, bounce_buf_len)
  412. ),
  413. TP_fast_assign(
  414. __entry->ring = ring;
  415. __entry->type = ring->type;
  416. __entry->num_segs = ring->num_segs;
  417. __entry->stream_id = ring->stream_id;
  418. __entry->enq_seg = ring->enq_seg->dma;
  419. __entry->deq_seg = ring->deq_seg->dma;
  420. __entry->cycle_state = ring->cycle_state;
  421. __entry->num_trbs_free = ring->num_trbs_free;
  422. __entry->bounce_buf_len = ring->bounce_buf_len;
  423. __entry->enq = xhci_trb_virt_to_dma(ring->enq_seg, ring->enqueue);
  424. __entry->deq = xhci_trb_virt_to_dma(ring->deq_seg, ring->dequeue);
  425. ),
  426. TP_printk("%s %p: enq %pad(%pad) deq %pad(%pad) segs %d stream %d free_trbs %d bounce %d cycle %d",
  427. xhci_ring_type_string(__entry->type), __entry->ring,
  428. &__entry->enq, &__entry->enq_seg,
  429. &__entry->deq, &__entry->deq_seg,
  430. __entry->num_segs,
  431. __entry->stream_id,
  432. __entry->num_trbs_free,
  433. __entry->bounce_buf_len,
  434. __entry->cycle_state
  435. )
  436. );
  437. DEFINE_EVENT(xhci_log_ring, xhci_ring_alloc,
  438. TP_PROTO(struct xhci_ring *ring),
  439. TP_ARGS(ring)
  440. );
  441. DEFINE_EVENT(xhci_log_ring, xhci_ring_free,
  442. TP_PROTO(struct xhci_ring *ring),
  443. TP_ARGS(ring)
  444. );
  445. DEFINE_EVENT(xhci_log_ring, xhci_ring_expansion,
  446. TP_PROTO(struct xhci_ring *ring),
  447. TP_ARGS(ring)
  448. );
  449. DEFINE_EVENT(xhci_log_ring, xhci_inc_enq,
  450. TP_PROTO(struct xhci_ring *ring),
  451. TP_ARGS(ring)
  452. );
  453. DEFINE_EVENT(xhci_log_ring, xhci_inc_deq,
  454. TP_PROTO(struct xhci_ring *ring),
  455. TP_ARGS(ring)
  456. );
  457. DECLARE_EVENT_CLASS(xhci_log_portsc,
  458. TP_PROTO(u32 portnum, u32 portsc),
  459. TP_ARGS(portnum, portsc),
  460. TP_STRUCT__entry(
  461. __field(u32, portnum)
  462. __field(u32, portsc)
  463. __dynamic_array(char, str, XHCI_MSG_MAX)
  464. ),
  465. TP_fast_assign(
  466. __entry->portnum = portnum;
  467. __entry->portsc = portsc;
  468. ),
  469. TP_printk("port-%d: %s",
  470. __entry->portnum,
  471. xhci_decode_portsc(__get_str(str), __entry->portsc)
  472. )
  473. );
  474. DEFINE_EVENT(xhci_log_portsc, xhci_handle_port_status,
  475. TP_PROTO(u32 portnum, u32 portsc),
  476. TP_ARGS(portnum, portsc)
  477. );
  478. DEFINE_EVENT(xhci_log_portsc, xhci_get_port_status,
  479. TP_PROTO(u32 portnum, u32 portsc),
  480. TP_ARGS(portnum, portsc)
  481. );
  482. DEFINE_EVENT(xhci_log_portsc, xhci_hub_status_data,
  483. TP_PROTO(u32 portnum, u32 portsc),
  484. TP_ARGS(portnum, portsc)
  485. );
  486. DECLARE_EVENT_CLASS(xhci_log_doorbell,
  487. TP_PROTO(u32 slot, u32 doorbell),
  488. TP_ARGS(slot, doorbell),
  489. TP_STRUCT__entry(
  490. __field(u32, slot)
  491. __field(u32, doorbell)
  492. __dynamic_array(char, str, XHCI_MSG_MAX)
  493. ),
  494. TP_fast_assign(
  495. __entry->slot = slot;
  496. __entry->doorbell = doorbell;
  497. ),
  498. TP_printk("Ring doorbell for %s",
  499. xhci_decode_doorbell(__get_str(str), __entry->slot, __entry->doorbell)
  500. )
  501. );
  502. DEFINE_EVENT(xhci_log_doorbell, xhci_ring_ep_doorbell,
  503. TP_PROTO(u32 slot, u32 doorbell),
  504. TP_ARGS(slot, doorbell)
  505. );
  506. DEFINE_EVENT(xhci_log_doorbell, xhci_ring_host_doorbell,
  507. TP_PROTO(u32 slot, u32 doorbell),
  508. TP_ARGS(slot, doorbell)
  509. );
  510. DECLARE_EVENT_CLASS(xhci_dbc_log_request,
  511. TP_PROTO(struct dbc_request *req),
  512. TP_ARGS(req),
  513. TP_STRUCT__entry(
  514. __field(struct dbc_request *, req)
  515. __field(bool, dir)
  516. __field(unsigned int, actual)
  517. __field(unsigned int, length)
  518. __field(int, status)
  519. ),
  520. TP_fast_assign(
  521. __entry->req = req;
  522. __entry->dir = req->direction;
  523. __entry->actual = req->actual;
  524. __entry->length = req->length;
  525. __entry->status = req->status;
  526. ),
  527. TP_printk("%s: req %p length %u/%u ==> %d",
  528. __entry->dir ? "bulk-in" : "bulk-out",
  529. __entry->req, __entry->actual,
  530. __entry->length, __entry->status
  531. )
  532. );
  533. DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_alloc_request,
  534. TP_PROTO(struct dbc_request *req),
  535. TP_ARGS(req)
  536. );
  537. DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_free_request,
  538. TP_PROTO(struct dbc_request *req),
  539. TP_ARGS(req)
  540. );
  541. DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_queue_request,
  542. TP_PROTO(struct dbc_request *req),
  543. TP_ARGS(req)
  544. );
  545. DEFINE_EVENT(xhci_dbc_log_request, xhci_dbc_giveback_request,
  546. TP_PROTO(struct dbc_request *req),
  547. TP_ARGS(req)
  548. );
  549. #endif /* __XHCI_TRACE_H */
  550. /* this part must be outside header guard */
  551. #undef TRACE_INCLUDE_PATH
  552. #define TRACE_INCLUDE_PATH .
  553. #undef TRACE_INCLUDE_FILE
  554. #define TRACE_INCLUDE_FILE xhci-trace
  555. #include <trace/define_trace.h>