adreno_a6xx_ringbuffer.c 16 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578
  1. // SPDX-License-Identifier: GPL-2.0-only
  2. /*
  3. * Copyright (c) 2020-2021, The Linux Foundation. All rights reserved.
  4. * Copyright (c) 2022-2023 Qualcomm Innovation Center, Inc. All rights reserved.
  5. */
  6. #include "adreno.h"
  7. #include "adreno_a6xx.h"
  8. #include "adreno_pm4types.h"
  9. #include "adreno_ringbuffer.h"
  10. #include "adreno_trace.h"
  11. #include "kgsl_trace.h"
  12. static int a6xx_rb_pagetable_switch(struct adreno_device *adreno_dev,
  13. struct adreno_ringbuffer *rb, struct adreno_context *drawctxt,
  14. struct kgsl_pagetable *pagetable, u32 *cmds)
  15. {
  16. struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
  17. u64 ttbr0 = kgsl_mmu_pagetable_get_ttbr0(pagetable);
  18. int count = 0;
  19. u32 id = drawctxt ? drawctxt->base.id : 0;
  20. if (pagetable == device->mmu.defaultpagetable)
  21. return 0;
  22. cmds[count++] = cp_type7_packet(CP_SMMU_TABLE_UPDATE, 3);
  23. cmds[count++] = lower_32_bits(ttbr0);
  24. cmds[count++] = upper_32_bits(ttbr0);
  25. cmds[count++] = id;
  26. if (!ADRENO_FEATURE(adreno_dev, ADRENO_APRIV)) {
  27. cmds[count++] = cp_type7_packet(CP_WAIT_FOR_IDLE, 0);
  28. cmds[count++] = cp_type7_packet(CP_WAIT_FOR_ME, 0);
  29. cmds[count++] = cp_type4_packet(A6XX_CP_MISC_CNTL, 1);
  30. cmds[count++] = 1;
  31. }
  32. cmds[count++] = cp_type7_packet(CP_MEM_WRITE, 5);
  33. cmds[count++] = lower_32_bits(SCRATCH_RB_GPU_ADDR(device,
  34. rb->id, ttbr0));
  35. cmds[count++] = upper_32_bits(SCRATCH_RB_GPU_ADDR(device,
  36. rb->id, ttbr0));
  37. cmds[count++] = lower_32_bits(ttbr0);
  38. cmds[count++] = upper_32_bits(ttbr0);
  39. cmds[count++] = id;
  40. if (!ADRENO_FEATURE(adreno_dev, ADRENO_APRIV)) {
  41. cmds[count++] = cp_type7_packet(CP_WAIT_FOR_IDLE, 0);
  42. cmds[count++] = cp_type7_packet(CP_WAIT_FOR_ME, 0);
  43. cmds[count++] = cp_type4_packet(A6XX_CP_MISC_CNTL, 1);
  44. cmds[count++] = 0;
  45. }
  46. return count;
  47. }
  48. static int a6xx_rb_context_switch(struct adreno_device *adreno_dev,
  49. struct adreno_ringbuffer *rb,
  50. struct adreno_context *drawctxt)
  51. {
  52. struct kgsl_pagetable *pagetable =
  53. adreno_drawctxt_get_pagetable(drawctxt);
  54. struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
  55. int count = 0;
  56. u32 cmds[41];
  57. if (adreno_drawctxt_get_pagetable(rb->drawctxt_active) != pagetable) {
  58. /* Clear performance counters during context switches */
  59. if (!adreno_dev->perfcounter) {
  60. cmds[count++] = cp_type4_packet(A6XX_RBBM_PERFCTR_SRAM_INIT_CMD, 1);
  61. cmds[count++] = 0x1;
  62. }
  63. count += a6xx_rb_pagetable_switch(adreno_dev, rb, drawctxt,
  64. pagetable, &cmds[count]);
  65. /* Wait for performance counter clear to finish */
  66. if (!adreno_dev->perfcounter) {
  67. cmds[count++] = cp_type7_packet(CP_WAIT_REG_MEM, 6);
  68. cmds[count++] = 0x3;
  69. cmds[count++] = A6XX_RBBM_PERFCTR_SRAM_INIT_STATUS;
  70. cmds[count++] = 0x0;
  71. cmds[count++] = 0x1;
  72. cmds[count++] = 0x1;
  73. cmds[count++] = 0x0;
  74. }
  75. }
  76. cmds[count++] = cp_type7_packet(CP_NOP, 1);
  77. cmds[count++] = CONTEXT_TO_MEM_IDENTIFIER;
  78. cmds[count++] = cp_type7_packet(CP_MEM_WRITE, 3);
  79. cmds[count++] = lower_32_bits(MEMSTORE_RB_GPU_ADDR(device, rb,
  80. current_context));
  81. cmds[count++] = upper_32_bits(MEMSTORE_RB_GPU_ADDR(device, rb,
  82. current_context));
  83. cmds[count++] = drawctxt->base.id;
  84. cmds[count++] = cp_type7_packet(CP_MEM_WRITE, 3);
  85. cmds[count++] = lower_32_bits(MEMSTORE_ID_GPU_ADDR(device,
  86. KGSL_MEMSTORE_GLOBAL, current_context));
  87. cmds[count++] = upper_32_bits(MEMSTORE_ID_GPU_ADDR(device,
  88. KGSL_MEMSTORE_GLOBAL, current_context));
  89. cmds[count++] = drawctxt->base.id;
  90. cmds[count++] = cp_type7_packet(CP_EVENT_WRITE, 1);
  91. cmds[count++] = 0x31;
  92. return a6xx_ringbuffer_addcmds(adreno_dev, rb, NULL, F_NOTPROTECTED,
  93. cmds, count, 0, NULL);
  94. }
  95. #define RB_SOPTIMESTAMP(device, rb) \
  96. MEMSTORE_RB_GPU_ADDR(device, rb, soptimestamp)
  97. #define CTXT_SOPTIMESTAMP(device, drawctxt) \
  98. MEMSTORE_ID_GPU_ADDR(device, (drawctxt)->base.id, soptimestamp)
  99. #define RB_EOPTIMESTAMP(device, rb) \
  100. MEMSTORE_RB_GPU_ADDR(device, rb, eoptimestamp)
  101. #define CTXT_EOPTIMESTAMP(device, drawctxt) \
  102. MEMSTORE_ID_GPU_ADDR(device, (drawctxt)->base.id, eoptimestamp)
  103. int a6xx_ringbuffer_submit(struct adreno_ringbuffer *rb,
  104. struct adreno_submit_time *time, bool sync)
  105. {
  106. struct adreno_device *adreno_dev = ADRENO_RB_DEVICE(rb);
  107. struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
  108. int ret = 0;
  109. unsigned long flags;
  110. adreno_get_submit_time(adreno_dev, rb, time);
  111. adreno_profile_submit_time(time);
  112. if (sync && !ADRENO_FEATURE(adreno_dev, ADRENO_APRIV)) {
  113. u32 *cmds = adreno_ringbuffer_allocspace(rb, 3);
  114. if (IS_ERR(cmds))
  115. return PTR_ERR(cmds);
  116. cmds[0] = cp_type7_packet(CP_WHERE_AM_I, 2);
  117. cmds[1] = lower_32_bits(SCRATCH_RB_GPU_ADDR(device, rb->id,
  118. rptr));
  119. cmds[2] = upper_32_bits(SCRATCH_RB_GPU_ADDR(device, rb->id,
  120. rptr));
  121. }
  122. spin_lock_irqsave(&rb->preempt_lock, flags);
  123. if (adreno_in_preempt_state(adreno_dev, ADRENO_PREEMPT_NONE)) {
  124. if (adreno_dev->cur_rb == rb) {
  125. kgsl_pwrscale_busy(device);
  126. ret = a6xx_fenced_write(adreno_dev,
  127. A6XX_CP_RB_WPTR, rb->_wptr,
  128. FENCE_STATUS_WRITEDROPPED0_MASK);
  129. rb->skip_inline_wptr = false;
  130. }
  131. } else {
  132. if (adreno_dev->cur_rb == rb)
  133. rb->skip_inline_wptr = true;
  134. }
  135. rb->wptr = rb->_wptr;
  136. spin_unlock_irqrestore(&rb->preempt_lock, flags);
  137. if (ret) {
  138. /*
  139. * If WPTR update fails, take inline snapshot and trigger
  140. * recovery.
  141. */
  142. gmu_core_fault_snapshot(device);
  143. adreno_dispatcher_fault(adreno_dev,
  144. ADRENO_GMU_FAULT_SKIP_SNAPSHOT);
  145. }
  146. return ret;
  147. }
  148. int a6xx_ringbuffer_init(struct adreno_device *adreno_dev)
  149. {
  150. struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
  151. int i, ret;
  152. ret = adreno_allocate_global(device, &device->scratch, PAGE_SIZE,
  153. 0, 0, KGSL_MEMDESC_RANDOM | KGSL_MEMDESC_PRIVILEGED,
  154. "scratch");
  155. if (ret)
  156. return ret;
  157. adreno_dev->cur_rb = &(adreno_dev->ringbuffers[0]);
  158. if (!adreno_preemption_feature_set(adreno_dev)) {
  159. adreno_dev->num_ringbuffers = 1;
  160. return adreno_ringbuffer_setup(adreno_dev,
  161. &adreno_dev->ringbuffers[0], 0);
  162. }
  163. adreno_dev->num_ringbuffers = ARRAY_SIZE(adreno_dev->ringbuffers);
  164. for (i = 0; i < adreno_dev->num_ringbuffers; i++) {
  165. int ret;
  166. ret = adreno_ringbuffer_setup(adreno_dev,
  167. &adreno_dev->ringbuffers[i], i);
  168. if (ret)
  169. return ret;
  170. }
  171. timer_setup(&adreno_dev->preempt.timer, adreno_preemption_timer, 0);
  172. a6xx_preemption_init(adreno_dev);
  173. return 0;
  174. }
  175. #define A6XX_SUBMIT_MAX 79
  176. int a6xx_ringbuffer_addcmds(struct adreno_device *adreno_dev,
  177. struct adreno_ringbuffer *rb, struct adreno_context *drawctxt,
  178. u32 flags, u32 *in, u32 dwords, u32 timestamp,
  179. struct adreno_submit_time *time)
  180. {
  181. struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
  182. u32 size = A6XX_SUBMIT_MAX + dwords;
  183. u32 *cmds, index = 0;
  184. u64 profile_gpuaddr;
  185. u32 profile_dwords;
  186. if (adreno_drawctxt_detached(drawctxt))
  187. return -ENOENT;
  188. if (adreno_gpu_fault(adreno_dev) != 0)
  189. return -EPROTO;
  190. rb->timestamp++;
  191. if (drawctxt)
  192. drawctxt->internal_timestamp = rb->timestamp;
  193. /*
  194. * if APRIV is enabled we assume all submissions are run with protected
  195. * mode off
  196. */
  197. if (ADRENO_FEATURE(adreno_dev, ADRENO_APRIV))
  198. flags &= ~F_NOTPROTECTED;
  199. cmds = adreno_ringbuffer_allocspace(rb, size);
  200. if (IS_ERR(cmds))
  201. return PTR_ERR(cmds);
  202. /* Identify the start of a command */
  203. cmds[index++] = cp_type7_packet(CP_NOP, 1);
  204. cmds[index++] = drawctxt ? CMD_IDENTIFIER : CMD_INTERNAL_IDENTIFIER;
  205. /* This is 25 dwords when drawctxt is not NULL and perfcounter needs to be zapped */
  206. index += a6xx_preemption_pre_ibsubmit(adreno_dev, rb, drawctxt,
  207. &cmds[index]);
  208. cmds[index++] = cp_type7_packet(CP_SET_MARKER, 1);
  209. cmds[index++] = 0x101; /* IFPC disable */
  210. profile_gpuaddr = adreno_profile_preib_processing(adreno_dev,
  211. drawctxt, &profile_dwords);
  212. if (profile_gpuaddr) {
  213. cmds[index++] = cp_type7_packet(CP_INDIRECT_BUFFER_PFE, 3);
  214. cmds[index++] = lower_32_bits(profile_gpuaddr);
  215. cmds[index++] = upper_32_bits(profile_gpuaddr);
  216. cmds[index++] = profile_dwords;
  217. }
  218. if (drawctxt) {
  219. cmds[index++] = cp_type7_packet(CP_MEM_WRITE, 3);
  220. cmds[index++] = lower_32_bits(CTXT_SOPTIMESTAMP(device,
  221. drawctxt));
  222. cmds[index++] = upper_32_bits(CTXT_SOPTIMESTAMP(device,
  223. drawctxt));
  224. cmds[index++] = timestamp;
  225. }
  226. cmds[index++] = cp_type7_packet(CP_MEM_WRITE, 3);
  227. cmds[index++] = lower_32_bits(RB_SOPTIMESTAMP(device, rb));
  228. cmds[index++] = upper_32_bits(RB_SOPTIMESTAMP(device, rb));
  229. cmds[index++] = rb->timestamp;
  230. if (IS_SECURE(flags)) {
  231. cmds[index++] = cp_type7_packet(CP_SET_SECURE_MODE, 1);
  232. cmds[index++] = 1;
  233. }
  234. if (IS_NOTPROTECTED(flags)) {
  235. cmds[index++] = cp_type7_packet(CP_SET_PROTECTED_MODE, 1);
  236. cmds[index++] = 0;
  237. }
  238. memcpy(&cmds[index], in, dwords << 2);
  239. index += dwords;
  240. if (IS_NOTPROTECTED(flags)) {
  241. cmds[index++] = cp_type7_packet(CP_SET_PROTECTED_MODE, 1);
  242. cmds[index++] = 1;
  243. }
  244. profile_gpuaddr = adreno_profile_postib_processing(adreno_dev,
  245. drawctxt, &dwords);
  246. if (profile_gpuaddr) {
  247. cmds[index++] = cp_type7_packet(CP_INDIRECT_BUFFER_PFE, 3);
  248. cmds[index++] = lower_32_bits(profile_gpuaddr);
  249. cmds[index++] = upper_32_bits(profile_gpuaddr);
  250. cmds[index++] = profile_dwords;
  251. }
  252. if (test_bit(KGSL_FT_PAGEFAULT_GPUHALT_ENABLE, &device->mmu.pfpolicy))
  253. cmds[index++] = cp_type7_packet(CP_WAIT_MEM_WRITES, 0);
  254. /*
  255. * If this is an internal command, just write the ringbuffer timestamp,
  256. * otherwise, write both
  257. */
  258. if (!drawctxt) {
  259. cmds[index++] = cp_type7_packet(CP_EVENT_WRITE, 4);
  260. cmds[index++] = CACHE_FLUSH_TS | (1 << 31);
  261. cmds[index++] = lower_32_bits(RB_EOPTIMESTAMP(device, rb));
  262. cmds[index++] = upper_32_bits(RB_EOPTIMESTAMP(device, rb));
  263. cmds[index++] = rb->timestamp;
  264. } else {
  265. cmds[index++] = cp_type7_packet(CP_EVENT_WRITE, 4);
  266. cmds[index++] = CACHE_FLUSH_TS | (1 << 31);
  267. cmds[index++] = lower_32_bits(CTXT_EOPTIMESTAMP(device,
  268. drawctxt));
  269. cmds[index++] = upper_32_bits(CTXT_EOPTIMESTAMP(device,
  270. drawctxt));
  271. cmds[index++] = timestamp;
  272. cmds[index++] = cp_type7_packet(CP_EVENT_WRITE, 4);
  273. cmds[index++] = CACHE_FLUSH_TS;
  274. cmds[index++] = lower_32_bits(RB_EOPTIMESTAMP(device, rb));
  275. cmds[index++] = upper_32_bits(RB_EOPTIMESTAMP(device, rb));
  276. cmds[index++] = rb->timestamp;
  277. }
  278. cmds[index++] = cp_type7_packet(CP_SET_MARKER, 1);
  279. cmds[index++] = 0x100; /* IFPC enable */
  280. if (IS_WFI(flags))
  281. cmds[index++] = cp_type7_packet(CP_WAIT_FOR_IDLE, 0);
  282. if (IS_SECURE(flags)) {
  283. cmds[index++] = cp_type7_packet(CP_SET_SECURE_MODE, 1);
  284. cmds[index++] = 0;
  285. }
  286. /* 10 dwords */
  287. index += a6xx_preemption_post_ibsubmit(adreno_dev, &cmds[index]);
  288. /* Adjust the thing for the number of dwords we actually wrote */
  289. rb->_wptr -= (size - index);
  290. return a6xx_ringbuffer_submit(rb, time,
  291. !adreno_is_preemption_enabled(adreno_dev));
  292. }
  293. static u32 a6xx_get_alwayson_counter(u32 *cmds, u64 gpuaddr)
  294. {
  295. cmds[0] = cp_type7_packet(CP_REG_TO_MEM, 3);
  296. cmds[1] = A6XX_CP_ALWAYS_ON_COUNTER_LO | (1 << 30) | (2 << 18);
  297. cmds[2] = lower_32_bits(gpuaddr);
  298. cmds[3] = upper_32_bits(gpuaddr);
  299. return 4;
  300. }
  301. static u32 a6xx_get_alwayson_context(u32 *cmds, u64 gpuaddr)
  302. {
  303. cmds[0] = cp_type7_packet(CP_REG_TO_MEM, 3);
  304. cmds[1] = A6XX_CP_ALWAYS_ON_CONTEXT_LO | (1 << 30) | (2 << 18);
  305. cmds[2] = lower_32_bits(gpuaddr);
  306. cmds[3] = upper_32_bits(gpuaddr);
  307. return 4;
  308. }
  309. #define PROFILE_IB_DWORDS 4
  310. #define PROFILE_IB_SLOTS (PAGE_SIZE / (PROFILE_IB_DWORDS << 2))
  311. static u64 a6xx_get_user_profiling_ib(struct adreno_ringbuffer *rb,
  312. struct kgsl_drawobj_cmd *cmdobj, u32 target_offset, u32 *cmds)
  313. {
  314. u32 offset, *ib, dwords;
  315. u64 gpuaddr;
  316. if (IS_ERR(rb->profile_desc))
  317. return 0;
  318. offset = rb->profile_index * (PROFILE_IB_DWORDS << 2);
  319. ib = rb->profile_desc->hostptr + offset;
  320. gpuaddr = rb->profile_desc->gpuaddr + offset;
  321. dwords = a6xx_get_alwayson_counter(ib,
  322. cmdobj->profiling_buffer_gpuaddr + target_offset);
  323. cmds[0] = cp_type7_packet(CP_INDIRECT_BUFFER_PFE, 3);
  324. cmds[1] = lower_32_bits(gpuaddr);
  325. cmds[2] = upper_32_bits(gpuaddr);
  326. cmds[3] = dwords;
  327. rb->profile_index = (rb->profile_index + 1) % PROFILE_IB_SLOTS;
  328. return 4;
  329. }
  330. static int a6xx_drawctxt_switch(struct adreno_device *adreno_dev,
  331. struct adreno_ringbuffer *rb,
  332. struct adreno_context *drawctxt)
  333. {
  334. struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
  335. int ret;
  336. if (rb->drawctxt_active == drawctxt)
  337. return 0;
  338. if (kgsl_context_detached(&drawctxt->base))
  339. return -ENOENT;
  340. if (!_kgsl_context_get(&drawctxt->base))
  341. return -ENOENT;
  342. ret = a6xx_rb_context_switch(adreno_dev, rb, drawctxt);
  343. if (ret) {
  344. kgsl_context_put(&drawctxt->base);
  345. return ret;
  346. }
  347. trace_adreno_drawctxt_switch(rb, drawctxt);
  348. /* Release the current drawctxt as soon as the new one is switched */
  349. adreno_put_drawctxt_on_timestamp(device, rb->drawctxt_active,
  350. rb, rb->timestamp);
  351. rb->drawctxt_active = drawctxt;
  352. return 0;
  353. }
  354. #define A6XX_USER_PROFILE_IB(rb, cmdobj, cmds, field) \
  355. a6xx_get_user_profiling_ib((rb), (cmdobj), \
  356. offsetof(struct kgsl_drawobj_profiling_buffer, field), \
  357. (cmds))
  358. #define A6XX_KERNEL_PROFILE(dev, cmdobj, cmds, field) \
  359. a6xx_get_alwayson_counter((cmds), \
  360. (dev)->profile_buffer->gpuaddr + \
  361. ADRENO_DRAWOBJ_PROFILE_OFFSET((cmdobj)->profile_index, \
  362. field))
  363. #define A6XX_KERNEL_PROFILE_CONTEXT(dev, cmdobj, cmds, field) \
  364. a6xx_get_alwayson_context((cmds), \
  365. (dev)->profile_buffer->gpuaddr + \
  366. ADRENO_DRAWOBJ_PROFILE_OFFSET((cmdobj)->profile_index, \
  367. field))
  368. #define A6XX_COMMAND_DWORDS 40
  369. int a6xx_ringbuffer_submitcmd(struct adreno_device *adreno_dev,
  370. struct kgsl_drawobj_cmd *cmdobj, u32 flags,
  371. struct adreno_submit_time *time)
  372. {
  373. struct kgsl_device *device = KGSL_DEVICE(adreno_dev);
  374. struct kgsl_drawobj *drawobj = DRAWOBJ(cmdobj);
  375. struct adreno_context *drawctxt = ADRENO_CONTEXT(drawobj->context);
  376. struct adreno_ringbuffer *rb = drawctxt->rb;
  377. int ret = 0, numibs = 0, index = 0;
  378. u32 *cmds;
  379. /* Count the number of IBs (if we are not skipping) */
  380. if (!IS_SKIP(flags)) {
  381. struct list_head *tmp;
  382. list_for_each(tmp, &cmdobj->cmdlist)
  383. numibs++;
  384. }
  385. cmds = kvmalloc((A6XX_COMMAND_DWORDS + (numibs * 5)) << 2, GFP_KERNEL);
  386. if (!cmds) {
  387. ret = -ENOMEM;
  388. goto done;
  389. }
  390. cmds[index++] = cp_type7_packet(CP_NOP, 1);
  391. cmds[index++] = START_IB_IDENTIFIER;
  392. /* Kernel profiling: 8 dwords */
  393. if (IS_KERNEL_PROFILE(flags)) {
  394. index += A6XX_KERNEL_PROFILE(adreno_dev, cmdobj, &cmds[index],
  395. started);
  396. index += A6XX_KERNEL_PROFILE_CONTEXT(adreno_dev, cmdobj, &cmds[index],
  397. ctx_start);
  398. }
  399. /* User profiling: 4 dwords */
  400. if (IS_USER_PROFILE(flags))
  401. index += A6XX_USER_PROFILE_IB(rb, cmdobj, &cmds[index],
  402. gpu_ticks_submitted);
  403. if (numibs) {
  404. struct kgsl_memobj_node *ib;
  405. cmds[index++] = cp_type7_packet(CP_SET_MARKER, 1);
  406. cmds[index++] = 0x00d; /* IB1LIST start */
  407. list_for_each_entry(ib, &cmdobj->cmdlist, node) {
  408. if (ib->priv & MEMOBJ_SKIP ||
  409. (ib->flags & KGSL_CMDLIST_CTXTSWITCH_PREAMBLE
  410. && !IS_PREAMBLE(flags)))
  411. cmds[index++] = cp_type7_packet(CP_NOP, 4);
  412. cmds[index++] =
  413. cp_type7_packet(CP_INDIRECT_BUFFER_PFE, 3);
  414. cmds[index++] = lower_32_bits(ib->gpuaddr);
  415. cmds[index++] = upper_32_bits(ib->gpuaddr);
  416. /* Double check that IB_PRIV is never set */
  417. cmds[index++] = (ib->size >> 2) & 0xfffff;
  418. }
  419. cmds[index++] = cp_type7_packet(CP_SET_MARKER, 1);
  420. cmds[index++] = 0x00e; /* IB1LIST end */
  421. }
  422. /* CCU invalidate */
  423. cmds[index++] = cp_type7_packet(CP_EVENT_WRITE, 1);
  424. cmds[index++] = 24;
  425. cmds[index++] = cp_type7_packet(CP_EVENT_WRITE, 1);
  426. cmds[index++] = 25;
  427. /* 8 dwords */
  428. if (IS_KERNEL_PROFILE(flags)) {
  429. index += A6XX_KERNEL_PROFILE(adreno_dev, cmdobj, &cmds[index],
  430. retired);
  431. index += A6XX_KERNEL_PROFILE_CONTEXT(adreno_dev, cmdobj, &cmds[index],
  432. ctx_end);
  433. }
  434. /* 4 dwords */
  435. if (IS_USER_PROFILE(flags))
  436. index += A6XX_USER_PROFILE_IB(rb, cmdobj, &cmds[index],
  437. gpu_ticks_retired);
  438. cmds[index++] = cp_type7_packet(CP_NOP, 1);
  439. cmds[index++] = END_IB_IDENTIFIER;
  440. ret = a6xx_drawctxt_switch(adreno_dev, rb, drawctxt);
  441. /*
  442. * In the unlikely event of an error in the drawctxt switch,
  443. * treat it like a hang
  444. */
  445. if (ret) {
  446. /*
  447. * It is "normal" to get a -ENOSPC or a -ENOENT. Don't log it,
  448. * the upper layers know how to handle it
  449. */
  450. if (ret != -ENOSPC && ret != -ENOENT)
  451. dev_err(device->dev,
  452. "Unable to switch draw context: %d\n",
  453. ret);
  454. goto done;
  455. }
  456. adreno_drawobj_set_constraint(device, drawobj);
  457. ret = a6xx_ringbuffer_addcmds(adreno_dev, drawctxt->rb, drawctxt,
  458. flags, cmds, index, drawobj->timestamp, time);
  459. done:
  460. trace_kgsl_issueibcmds(device, drawctxt->base.id, numibs,
  461. drawobj->timestamp, drawobj->flags, ret, drawctxt->type);
  462. kvfree(cmds);
  463. return ret;
  464. }