virtgpu_drv.h 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487
  1. /*
  2. * Copyright (C) 2015 Red Hat, Inc.
  3. * All Rights Reserved.
  4. *
  5. * Permission is hereby granted, free of charge, to any person obtaining
  6. * a copy of this software and associated documentation files (the
  7. * "Software"), to deal in the Software without restriction, including
  8. * without limitation the rights to use, copy, modify, merge, publish,
  9. * distribute, sublicense, and/or sell copies of the Software, and to
  10. * permit persons to whom the Software is furnished to do so, subject to
  11. * the following conditions:
  12. *
  13. * The above copyright notice and this permission notice (including the
  14. * next paragraph) shall be included in all copies or substantial
  15. * portions of the Software.
  16. *
  17. * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
  18. * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
  19. * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
  20. * IN NO EVENT SHALL THE COPYRIGHT OWNER(S) AND/OR ITS SUPPLIERS BE
  21. * LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION
  22. * OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION
  23. * WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.
  24. */
  25. #ifndef VIRTIO_DRV_H
  26. #define VIRTIO_DRV_H
  27. #include <linux/dma-direction.h>
  28. #include <linux/virtio.h>
  29. #include <linux/virtio_ids.h>
  30. #include <linux/virtio_config.h>
  31. #include <linux/virtio_gpu.h>
  32. #include <drm/drm_atomic.h>
  33. #include <drm/drm_drv.h>
  34. #include <drm/drm_encoder.h>
  35. #include <drm/drm_fb_helper.h>
  36. #include <drm/drm_fourcc.h>
  37. #include <drm/drm_framebuffer.h>
  38. #include <drm/drm_gem.h>
  39. #include <drm/drm_gem_shmem_helper.h>
  40. #include <drm/drm_ioctl.h>
  41. #include <drm/drm_probe_helper.h>
  42. #include <drm/virtgpu_drm.h>
  43. #define DRIVER_NAME "virtio_gpu"
  44. #define DRIVER_DESC "virtio GPU"
  45. #define DRIVER_DATE "0"
  46. #define DRIVER_MAJOR 0
  47. #define DRIVER_MINOR 1
  48. #define DRIVER_PATCHLEVEL 0
  49. #define STATE_INITIALIZING 0
  50. #define STATE_OK 1
  51. #define STATE_ERR 2
  52. #define MAX_CAPSET_ID 63
  53. #define MAX_RINGS 64
  54. struct virtio_gpu_object_params {
  55. unsigned long size;
  56. bool dumb;
  57. /* 3d */
  58. bool virgl;
  59. bool blob;
  60. /* classic resources only */
  61. uint32_t format;
  62. uint32_t width;
  63. uint32_t height;
  64. uint32_t target;
  65. uint32_t bind;
  66. uint32_t depth;
  67. uint32_t array_size;
  68. uint32_t last_level;
  69. uint32_t nr_samples;
  70. uint32_t flags;
  71. /* blob resources only */
  72. uint32_t ctx_id;
  73. uint32_t blob_mem;
  74. uint32_t blob_flags;
  75. uint64_t blob_id;
  76. };
  77. struct virtio_gpu_object {
  78. struct drm_gem_shmem_object base;
  79. uint32_t hw_res_handle;
  80. bool dumb;
  81. bool created;
  82. bool host3d_blob, guest_blob;
  83. uint32_t blob_mem, blob_flags;
  84. int uuid_state;
  85. uuid_t uuid;
  86. };
  87. #define gem_to_virtio_gpu_obj(gobj) \
  88. container_of((gobj), struct virtio_gpu_object, base.base)
  89. struct virtio_gpu_object_shmem {
  90. struct virtio_gpu_object base;
  91. };
  92. struct virtio_gpu_object_vram {
  93. struct virtio_gpu_object base;
  94. uint32_t map_state;
  95. uint32_t map_info;
  96. struct drm_mm_node vram_node;
  97. };
  98. #define to_virtio_gpu_shmem(virtio_gpu_object) \
  99. container_of((virtio_gpu_object), struct virtio_gpu_object_shmem, base)
  100. #define to_virtio_gpu_vram(virtio_gpu_object) \
  101. container_of((virtio_gpu_object), struct virtio_gpu_object_vram, base)
  102. struct virtio_gpu_object_array {
  103. struct ww_acquire_ctx ticket;
  104. struct list_head next;
  105. u32 nents, total;
  106. struct drm_gem_object *objs[];
  107. };
  108. struct virtio_gpu_vbuffer;
  109. struct virtio_gpu_device;
  110. typedef void (*virtio_gpu_resp_cb)(struct virtio_gpu_device *vgdev,
  111. struct virtio_gpu_vbuffer *vbuf);
  112. struct virtio_gpu_fence_driver {
  113. atomic64_t last_fence_id;
  114. uint64_t current_fence_id;
  115. uint64_t context;
  116. struct list_head fences;
  117. spinlock_t lock;
  118. };
  119. struct virtio_gpu_fence_event {
  120. struct drm_pending_event base;
  121. struct drm_event event;
  122. };
  123. struct virtio_gpu_fence {
  124. struct dma_fence f;
  125. uint32_t ring_idx;
  126. uint64_t fence_id;
  127. bool emit_fence_info;
  128. struct virtio_gpu_fence_event *e;
  129. struct virtio_gpu_fence_driver *drv;
  130. struct list_head node;
  131. };
  132. struct virtio_gpu_vbuffer {
  133. char *buf;
  134. int size;
  135. void *data_buf;
  136. uint32_t data_size;
  137. char *resp_buf;
  138. int resp_size;
  139. virtio_gpu_resp_cb resp_cb;
  140. void *resp_cb_data;
  141. struct virtio_gpu_object_array *objs;
  142. struct list_head list;
  143. };
  144. struct virtio_gpu_output {
  145. int index;
  146. struct drm_crtc crtc;
  147. struct drm_connector conn;
  148. struct drm_encoder enc;
  149. struct virtio_gpu_display_one info;
  150. struct virtio_gpu_update_cursor cursor;
  151. struct edid *edid;
  152. int cur_x;
  153. int cur_y;
  154. bool needs_modeset;
  155. };
  156. #define drm_crtc_to_virtio_gpu_output(x) \
  157. container_of(x, struct virtio_gpu_output, crtc)
  158. struct virtio_gpu_framebuffer {
  159. struct drm_framebuffer base;
  160. struct virtio_gpu_fence *fence;
  161. };
  162. #define to_virtio_gpu_framebuffer(x) \
  163. container_of(x, struct virtio_gpu_framebuffer, base)
  164. struct virtio_gpu_queue {
  165. struct virtqueue *vq;
  166. spinlock_t qlock;
  167. wait_queue_head_t ack_queue;
  168. struct work_struct dequeue_work;
  169. };
  170. struct virtio_gpu_drv_capset {
  171. uint32_t id;
  172. uint32_t max_version;
  173. uint32_t max_size;
  174. };
  175. struct virtio_gpu_drv_cap_cache {
  176. struct list_head head;
  177. void *caps_cache;
  178. uint32_t id;
  179. uint32_t version;
  180. uint32_t size;
  181. atomic_t is_valid;
  182. };
  183. struct virtio_gpu_device {
  184. struct drm_device *ddev;
  185. struct virtio_device *vdev;
  186. struct virtio_gpu_output outputs[VIRTIO_GPU_MAX_SCANOUTS];
  187. uint32_t num_scanouts;
  188. struct virtio_gpu_queue ctrlq;
  189. struct virtio_gpu_queue cursorq;
  190. struct kmem_cache *vbufs;
  191. atomic_t pending_commands;
  192. struct ida resource_ida;
  193. wait_queue_head_t resp_wq;
  194. /* current display info */
  195. spinlock_t display_info_lock;
  196. bool display_info_pending;
  197. struct virtio_gpu_fence_driver fence_drv;
  198. struct ida ctx_id_ida;
  199. bool has_virgl_3d;
  200. bool has_edid;
  201. bool has_indirect;
  202. bool has_resource_assign_uuid;
  203. bool has_resource_blob;
  204. bool has_host_visible;
  205. bool has_context_init;
  206. struct virtio_shm_region host_visible_region;
  207. struct drm_mm host_visible_mm;
  208. struct work_struct config_changed_work;
  209. struct work_struct obj_free_work;
  210. spinlock_t obj_free_lock;
  211. struct list_head obj_free_list;
  212. struct virtio_gpu_drv_capset *capsets;
  213. uint32_t num_capsets;
  214. uint64_t capset_id_mask;
  215. struct list_head cap_cache;
  216. /* protects uuid state when exporting */
  217. spinlock_t resource_export_lock;
  218. /* protects map state and host_visible_mm */
  219. spinlock_t host_visible_lock;
  220. };
  221. struct virtio_gpu_fpriv {
  222. uint32_t ctx_id;
  223. uint32_t context_init;
  224. bool context_created;
  225. uint32_t num_rings;
  226. uint64_t base_fence_ctx;
  227. uint64_t ring_idx_mask;
  228. struct mutex context_lock;
  229. };
  230. /* virtgpu_ioctl.c */
  231. #define DRM_VIRTIO_NUM_IOCTLS 12
  232. extern struct drm_ioctl_desc virtio_gpu_ioctls[DRM_VIRTIO_NUM_IOCTLS];
  233. void virtio_gpu_create_context(struct drm_device *dev, struct drm_file *file);
  234. /* virtgpu_kms.c */
  235. int virtio_gpu_init(struct virtio_device *vdev, struct drm_device *dev);
  236. void virtio_gpu_deinit(struct drm_device *dev);
  237. void virtio_gpu_release(struct drm_device *dev);
  238. int virtio_gpu_driver_open(struct drm_device *dev, struct drm_file *file);
  239. void virtio_gpu_driver_postclose(struct drm_device *dev, struct drm_file *file);
  240. /* virtgpu_gem.c */
  241. int virtio_gpu_gem_object_open(struct drm_gem_object *obj,
  242. struct drm_file *file);
  243. void virtio_gpu_gem_object_close(struct drm_gem_object *obj,
  244. struct drm_file *file);
  245. int virtio_gpu_mode_dumb_create(struct drm_file *file_priv,
  246. struct drm_device *dev,
  247. struct drm_mode_create_dumb *args);
  248. int virtio_gpu_mode_dumb_mmap(struct drm_file *file_priv,
  249. struct drm_device *dev,
  250. uint32_t handle, uint64_t *offset_p);
  251. struct virtio_gpu_object_array *virtio_gpu_array_alloc(u32 nents);
  252. struct virtio_gpu_object_array*
  253. virtio_gpu_array_from_handles(struct drm_file *drm_file, u32 *handles, u32 nents);
  254. void virtio_gpu_array_add_obj(struct virtio_gpu_object_array *objs,
  255. struct drm_gem_object *obj);
  256. int virtio_gpu_array_lock_resv(struct virtio_gpu_object_array *objs);
  257. void virtio_gpu_array_unlock_resv(struct virtio_gpu_object_array *objs);
  258. void virtio_gpu_array_add_fence(struct virtio_gpu_object_array *objs,
  259. struct dma_fence *fence);
  260. void virtio_gpu_array_put_free(struct virtio_gpu_object_array *objs);
  261. void virtio_gpu_array_put_free_delayed(struct virtio_gpu_device *vgdev,
  262. struct virtio_gpu_object_array *objs);
  263. void virtio_gpu_array_put_free_work(struct work_struct *work);
  264. /* virtgpu_vq.c */
  265. int virtio_gpu_alloc_vbufs(struct virtio_gpu_device *vgdev);
  266. void virtio_gpu_free_vbufs(struct virtio_gpu_device *vgdev);
  267. void virtio_gpu_cmd_create_resource(struct virtio_gpu_device *vgdev,
  268. struct virtio_gpu_object *bo,
  269. struct virtio_gpu_object_params *params,
  270. struct virtio_gpu_object_array *objs,
  271. struct virtio_gpu_fence *fence);
  272. void virtio_gpu_cmd_unref_resource(struct virtio_gpu_device *vgdev,
  273. struct virtio_gpu_object *bo);
  274. void virtio_gpu_cmd_transfer_to_host_2d(struct virtio_gpu_device *vgdev,
  275. uint64_t offset,
  276. uint32_t width, uint32_t height,
  277. uint32_t x, uint32_t y,
  278. struct virtio_gpu_object_array *objs,
  279. struct virtio_gpu_fence *fence);
  280. void virtio_gpu_cmd_resource_flush(struct virtio_gpu_device *vgdev,
  281. uint32_t resource_id,
  282. uint32_t x, uint32_t y,
  283. uint32_t width, uint32_t height,
  284. struct virtio_gpu_object_array *objs,
  285. struct virtio_gpu_fence *fence);
  286. void virtio_gpu_cmd_set_scanout(struct virtio_gpu_device *vgdev,
  287. uint32_t scanout_id, uint32_t resource_id,
  288. uint32_t width, uint32_t height,
  289. uint32_t x, uint32_t y);
  290. void virtio_gpu_object_attach(struct virtio_gpu_device *vgdev,
  291. struct virtio_gpu_object *obj,
  292. struct virtio_gpu_mem_entry *ents,
  293. unsigned int nents);
  294. int virtio_gpu_attach_status_page(struct virtio_gpu_device *vgdev);
  295. int virtio_gpu_detach_status_page(struct virtio_gpu_device *vgdev);
  296. void virtio_gpu_cursor_ping(struct virtio_gpu_device *vgdev,
  297. struct virtio_gpu_output *output);
  298. int virtio_gpu_cmd_get_display_info(struct virtio_gpu_device *vgdev);
  299. int virtio_gpu_cmd_get_capset_info(struct virtio_gpu_device *vgdev, int idx);
  300. int virtio_gpu_cmd_get_capset(struct virtio_gpu_device *vgdev,
  301. int idx, int version,
  302. struct virtio_gpu_drv_cap_cache **cache_p);
  303. int virtio_gpu_cmd_get_edids(struct virtio_gpu_device *vgdev);
  304. void virtio_gpu_cmd_context_create(struct virtio_gpu_device *vgdev, uint32_t id,
  305. uint32_t context_init, uint32_t nlen,
  306. const char *name);
  307. void virtio_gpu_cmd_context_destroy(struct virtio_gpu_device *vgdev,
  308. uint32_t id);
  309. void virtio_gpu_cmd_context_attach_resource(struct virtio_gpu_device *vgdev,
  310. uint32_t ctx_id,
  311. struct virtio_gpu_object_array *objs);
  312. void virtio_gpu_cmd_context_detach_resource(struct virtio_gpu_device *vgdev,
  313. uint32_t ctx_id,
  314. struct virtio_gpu_object_array *objs);
  315. void virtio_gpu_cmd_submit(struct virtio_gpu_device *vgdev,
  316. void *data, uint32_t data_size,
  317. uint32_t ctx_id,
  318. struct virtio_gpu_object_array *objs,
  319. struct virtio_gpu_fence *fence);
  320. void virtio_gpu_cmd_transfer_from_host_3d(struct virtio_gpu_device *vgdev,
  321. uint32_t ctx_id,
  322. uint64_t offset, uint32_t level,
  323. uint32_t stride,
  324. uint32_t layer_stride,
  325. struct drm_virtgpu_3d_box *box,
  326. struct virtio_gpu_object_array *objs,
  327. struct virtio_gpu_fence *fence);
  328. void virtio_gpu_cmd_transfer_to_host_3d(struct virtio_gpu_device *vgdev,
  329. uint32_t ctx_id,
  330. uint64_t offset, uint32_t level,
  331. uint32_t stride,
  332. uint32_t layer_stride,
  333. struct drm_virtgpu_3d_box *box,
  334. struct virtio_gpu_object_array *objs,
  335. struct virtio_gpu_fence *fence);
  336. void
  337. virtio_gpu_cmd_resource_create_3d(struct virtio_gpu_device *vgdev,
  338. struct virtio_gpu_object *bo,
  339. struct virtio_gpu_object_params *params,
  340. struct virtio_gpu_object_array *objs,
  341. struct virtio_gpu_fence *fence);
  342. void virtio_gpu_ctrl_ack(struct virtqueue *vq);
  343. void virtio_gpu_cursor_ack(struct virtqueue *vq);
  344. void virtio_gpu_fence_ack(struct virtqueue *vq);
  345. void virtio_gpu_dequeue_ctrl_func(struct work_struct *work);
  346. void virtio_gpu_dequeue_cursor_func(struct work_struct *work);
  347. void virtio_gpu_dequeue_fence_func(struct work_struct *work);
  348. void virtio_gpu_notify(struct virtio_gpu_device *vgdev);
  349. int
  350. virtio_gpu_cmd_resource_assign_uuid(struct virtio_gpu_device *vgdev,
  351. struct virtio_gpu_object_array *objs);
  352. int virtio_gpu_cmd_map(struct virtio_gpu_device *vgdev,
  353. struct virtio_gpu_object_array *objs, uint64_t offset);
  354. void virtio_gpu_cmd_unmap(struct virtio_gpu_device *vgdev,
  355. struct virtio_gpu_object *bo);
  356. void
  357. virtio_gpu_cmd_resource_create_blob(struct virtio_gpu_device *vgdev,
  358. struct virtio_gpu_object *bo,
  359. struct virtio_gpu_object_params *params,
  360. struct virtio_gpu_mem_entry *ents,
  361. uint32_t nents);
  362. void
  363. virtio_gpu_cmd_set_scanout_blob(struct virtio_gpu_device *vgdev,
  364. uint32_t scanout_id,
  365. struct virtio_gpu_object *bo,
  366. struct drm_framebuffer *fb,
  367. uint32_t width, uint32_t height,
  368. uint32_t x, uint32_t y);
  369. /* virtgpu_display.c */
  370. int virtio_gpu_modeset_init(struct virtio_gpu_device *vgdev);
  371. void virtio_gpu_modeset_fini(struct virtio_gpu_device *vgdev);
  372. /* virtgpu_plane.c */
  373. uint32_t virtio_gpu_translate_format(uint32_t drm_fourcc);
  374. struct drm_plane *virtio_gpu_plane_init(struct virtio_gpu_device *vgdev,
  375. enum drm_plane_type type,
  376. int index);
  377. /* virtgpu_fence.c */
  378. struct virtio_gpu_fence *virtio_gpu_fence_alloc(struct virtio_gpu_device *vgdev,
  379. uint64_t base_fence_ctx,
  380. uint32_t ring_idx);
  381. void virtio_gpu_fence_emit(struct virtio_gpu_device *vgdev,
  382. struct virtio_gpu_ctrl_hdr *cmd_hdr,
  383. struct virtio_gpu_fence *fence);
  384. void virtio_gpu_fence_event_process(struct virtio_gpu_device *vdev,
  385. u64 fence_id);
  386. /* virtgpu_object.c */
  387. void virtio_gpu_cleanup_object(struct virtio_gpu_object *bo);
  388. struct drm_gem_object *virtio_gpu_create_object(struct drm_device *dev,
  389. size_t size);
  390. int virtio_gpu_object_create(struct virtio_gpu_device *vgdev,
  391. struct virtio_gpu_object_params *params,
  392. struct virtio_gpu_object **bo_ptr,
  393. struct virtio_gpu_fence *fence);
  394. bool virtio_gpu_is_shmem(struct virtio_gpu_object *bo);
  395. int virtio_gpu_resource_id_get(struct virtio_gpu_device *vgdev,
  396. uint32_t *resid);
  397. /* virtgpu_prime.c */
  398. int virtio_gpu_resource_assign_uuid(struct virtio_gpu_device *vgdev,
  399. struct virtio_gpu_object *bo);
  400. struct dma_buf *virtgpu_gem_prime_export(struct drm_gem_object *obj,
  401. int flags);
  402. struct drm_gem_object *virtgpu_gem_prime_import(struct drm_device *dev,
  403. struct dma_buf *buf);
  404. int virtgpu_gem_prime_get_uuid(struct drm_gem_object *obj,
  405. uuid_t *uuid);
  406. struct drm_gem_object *virtgpu_gem_prime_import_sg_table(
  407. struct drm_device *dev, struct dma_buf_attachment *attach,
  408. struct sg_table *sgt);
  409. /* virtgpu_debugfs.c */
  410. void virtio_gpu_debugfs_init(struct drm_minor *minor);
  411. /* virtgpu_vram.c */
  412. bool virtio_gpu_is_vram(struct virtio_gpu_object *bo);
  413. int virtio_gpu_vram_create(struct virtio_gpu_device *vgdev,
  414. struct virtio_gpu_object_params *params,
  415. struct virtio_gpu_object **bo_ptr);
  416. struct sg_table *virtio_gpu_vram_map_dma_buf(struct virtio_gpu_object *bo,
  417. struct device *dev,
  418. enum dma_data_direction dir);
  419. void virtio_gpu_vram_unmap_dma_buf(struct device *dev,
  420. struct sg_table *sgt,
  421. enum dma_data_direction dir);
  422. #endif