17d2ad4e1SGerd Hoffmann /* 27d2ad4e1SGerd Hoffmann * Virtio GPU Device 37d2ad4e1SGerd Hoffmann * 47d2ad4e1SGerd Hoffmann * Copyright Red Hat, Inc. 2013-2014 57d2ad4e1SGerd Hoffmann * 67d2ad4e1SGerd Hoffmann * Authors: 77d2ad4e1SGerd Hoffmann * Dave Airlie <airlied@redhat.com> 87d2ad4e1SGerd Hoffmann * Gerd Hoffmann <kraxel@redhat.com> 97d2ad4e1SGerd Hoffmann * 107d2ad4e1SGerd Hoffmann * This work is licensed under the terms of the GNU GPL, version 2 or later. 117d2ad4e1SGerd Hoffmann * See the COPYING file in the top-level directory. 127d2ad4e1SGerd Hoffmann */ 137d2ad4e1SGerd Hoffmann 147d2ad4e1SGerd Hoffmann #include "qemu/osdep.h" 157d2ad4e1SGerd Hoffmann #include "qemu/iov.h" 167d2ad4e1SGerd Hoffmann #include "trace.h" 177d2ad4e1SGerd Hoffmann #include "hw/virtio/virtio.h" 187d2ad4e1SGerd Hoffmann #include "hw/virtio/virtio-gpu.h" 197d2ad4e1SGerd Hoffmann 207d2ad4e1SGerd Hoffmann #include <virglrenderer.h> 217d2ad4e1SGerd Hoffmann 227d2ad4e1SGerd Hoffmann static struct virgl_renderer_callbacks virtio_gpu_3d_cbs; 237d2ad4e1SGerd Hoffmann 247d2ad4e1SGerd Hoffmann static void virgl_cmd_create_resource_2d(VirtIOGPU *g, 257d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 267d2ad4e1SGerd Hoffmann { 277d2ad4e1SGerd Hoffmann struct virtio_gpu_resource_create_2d c2d; 287d2ad4e1SGerd Hoffmann struct virgl_renderer_resource_create_args args; 297d2ad4e1SGerd Hoffmann 307d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(c2d); 317d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_create_2d(c2d.resource_id, c2d.format, 327d2ad4e1SGerd Hoffmann c2d.width, c2d.height); 337d2ad4e1SGerd Hoffmann 347d2ad4e1SGerd Hoffmann args.handle = c2d.resource_id; 357d2ad4e1SGerd Hoffmann args.target = 2; 367d2ad4e1SGerd Hoffmann args.format = c2d.format; 377d2ad4e1SGerd Hoffmann args.bind = (1 << 1); 387d2ad4e1SGerd Hoffmann args.width = c2d.width; 397d2ad4e1SGerd Hoffmann args.height = c2d.height; 407d2ad4e1SGerd Hoffmann args.depth = 1; 417d2ad4e1SGerd Hoffmann args.array_size = 1; 427d2ad4e1SGerd Hoffmann args.last_level = 0; 437d2ad4e1SGerd Hoffmann args.nr_samples = 0; 447d2ad4e1SGerd Hoffmann args.flags = VIRTIO_GPU_RESOURCE_FLAG_Y_0_TOP; 457d2ad4e1SGerd Hoffmann virgl_renderer_resource_create(&args, NULL, 0); 467d2ad4e1SGerd Hoffmann } 477d2ad4e1SGerd Hoffmann 487d2ad4e1SGerd Hoffmann static void virgl_cmd_create_resource_3d(VirtIOGPU *g, 497d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 507d2ad4e1SGerd Hoffmann { 517d2ad4e1SGerd Hoffmann struct virtio_gpu_resource_create_3d c3d; 527d2ad4e1SGerd Hoffmann struct virgl_renderer_resource_create_args args; 537d2ad4e1SGerd Hoffmann 547d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(c3d); 557d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_create_3d(c3d.resource_id, c3d.format, 567d2ad4e1SGerd Hoffmann c3d.width, c3d.height, c3d.depth); 577d2ad4e1SGerd Hoffmann 587d2ad4e1SGerd Hoffmann args.handle = c3d.resource_id; 597d2ad4e1SGerd Hoffmann args.target = c3d.target; 607d2ad4e1SGerd Hoffmann args.format = c3d.format; 617d2ad4e1SGerd Hoffmann args.bind = c3d.bind; 627d2ad4e1SGerd Hoffmann args.width = c3d.width; 637d2ad4e1SGerd Hoffmann args.height = c3d.height; 647d2ad4e1SGerd Hoffmann args.depth = c3d.depth; 657d2ad4e1SGerd Hoffmann args.array_size = c3d.array_size; 667d2ad4e1SGerd Hoffmann args.last_level = c3d.last_level; 677d2ad4e1SGerd Hoffmann args.nr_samples = c3d.nr_samples; 687d2ad4e1SGerd Hoffmann args.flags = c3d.flags; 697d2ad4e1SGerd Hoffmann virgl_renderer_resource_create(&args, NULL, 0); 707d2ad4e1SGerd Hoffmann } 717d2ad4e1SGerd Hoffmann 727d2ad4e1SGerd Hoffmann static void virgl_cmd_resource_unref(VirtIOGPU *g, 737d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 747d2ad4e1SGerd Hoffmann { 757d2ad4e1SGerd Hoffmann struct virtio_gpu_resource_unref unref; 767d2ad4e1SGerd Hoffmann struct iovec *res_iovs = NULL; 777d2ad4e1SGerd Hoffmann int num_iovs = 0; 787d2ad4e1SGerd Hoffmann 797d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(unref); 807d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_unref(unref.resource_id); 817d2ad4e1SGerd Hoffmann 827d2ad4e1SGerd Hoffmann virgl_renderer_resource_detach_iov(unref.resource_id, 837d2ad4e1SGerd Hoffmann &res_iovs, 847d2ad4e1SGerd Hoffmann &num_iovs); 857d2ad4e1SGerd Hoffmann if (res_iovs != NULL && num_iovs != 0) { 867d2ad4e1SGerd Hoffmann virtio_gpu_cleanup_mapping_iov(g, res_iovs, num_iovs); 877d2ad4e1SGerd Hoffmann } 887d2ad4e1SGerd Hoffmann virgl_renderer_resource_unref(unref.resource_id); 897d2ad4e1SGerd Hoffmann } 907d2ad4e1SGerd Hoffmann 917d2ad4e1SGerd Hoffmann static void virgl_cmd_context_create(VirtIOGPU *g, 927d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 937d2ad4e1SGerd Hoffmann { 947d2ad4e1SGerd Hoffmann struct virtio_gpu_ctx_create cc; 957d2ad4e1SGerd Hoffmann 967d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(cc); 977d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_ctx_create(cc.hdr.ctx_id, 987d2ad4e1SGerd Hoffmann cc.debug_name); 997d2ad4e1SGerd Hoffmann 1007d2ad4e1SGerd Hoffmann virgl_renderer_context_create(cc.hdr.ctx_id, cc.nlen, 1017d2ad4e1SGerd Hoffmann cc.debug_name); 1027d2ad4e1SGerd Hoffmann } 1037d2ad4e1SGerd Hoffmann 1047d2ad4e1SGerd Hoffmann static void virgl_cmd_context_destroy(VirtIOGPU *g, 1057d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 1067d2ad4e1SGerd Hoffmann { 1077d2ad4e1SGerd Hoffmann struct virtio_gpu_ctx_destroy cd; 1087d2ad4e1SGerd Hoffmann 1097d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(cd); 1107d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_ctx_destroy(cd.hdr.ctx_id); 1117d2ad4e1SGerd Hoffmann 1127d2ad4e1SGerd Hoffmann virgl_renderer_context_destroy(cd.hdr.ctx_id); 1137d2ad4e1SGerd Hoffmann } 1147d2ad4e1SGerd Hoffmann 1157d2ad4e1SGerd Hoffmann static void virtio_gpu_rect_update(VirtIOGPU *g, int idx, int x, int y, 1167d2ad4e1SGerd Hoffmann int width, int height) 1177d2ad4e1SGerd Hoffmann { 1187d2ad4e1SGerd Hoffmann if (!g->parent_obj.scanout[idx].con) { 1197d2ad4e1SGerd Hoffmann return; 1207d2ad4e1SGerd Hoffmann } 1217d2ad4e1SGerd Hoffmann 1227d2ad4e1SGerd Hoffmann dpy_gl_update(g->parent_obj.scanout[idx].con, x, y, width, height); 1237d2ad4e1SGerd Hoffmann } 1247d2ad4e1SGerd Hoffmann 1257d2ad4e1SGerd Hoffmann static void virgl_cmd_resource_flush(VirtIOGPU *g, 1267d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 1277d2ad4e1SGerd Hoffmann { 1287d2ad4e1SGerd Hoffmann struct virtio_gpu_resource_flush rf; 1297d2ad4e1SGerd Hoffmann int i; 1307d2ad4e1SGerd Hoffmann 1317d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(rf); 1327d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_flush(rf.resource_id, 1337d2ad4e1SGerd Hoffmann rf.r.width, rf.r.height, rf.r.x, rf.r.y); 1347d2ad4e1SGerd Hoffmann 1357d2ad4e1SGerd Hoffmann for (i = 0; i < g->parent_obj.conf.max_outputs; i++) { 1367d2ad4e1SGerd Hoffmann if (g->parent_obj.scanout[i].resource_id != rf.resource_id) { 1377d2ad4e1SGerd Hoffmann continue; 1387d2ad4e1SGerd Hoffmann } 1397d2ad4e1SGerd Hoffmann virtio_gpu_rect_update(g, i, rf.r.x, rf.r.y, rf.r.width, rf.r.height); 1407d2ad4e1SGerd Hoffmann } 1417d2ad4e1SGerd Hoffmann } 1427d2ad4e1SGerd Hoffmann 1437d2ad4e1SGerd Hoffmann static void virgl_cmd_set_scanout(VirtIOGPU *g, 1447d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 1457d2ad4e1SGerd Hoffmann { 1467d2ad4e1SGerd Hoffmann struct virtio_gpu_set_scanout ss; 1477d2ad4e1SGerd Hoffmann struct virgl_renderer_resource_info info; 1487d2ad4e1SGerd Hoffmann int ret; 1497d2ad4e1SGerd Hoffmann 1507d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(ss); 1517d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_set_scanout(ss.scanout_id, ss.resource_id, 1527d2ad4e1SGerd Hoffmann ss.r.width, ss.r.height, ss.r.x, ss.r.y); 1537d2ad4e1SGerd Hoffmann 1547d2ad4e1SGerd Hoffmann if (ss.scanout_id >= g->parent_obj.conf.max_outputs) { 1557d2ad4e1SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout id specified %d", 1567d2ad4e1SGerd Hoffmann __func__, ss.scanout_id); 1577d2ad4e1SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_SCANOUT_ID; 1587d2ad4e1SGerd Hoffmann return; 1597d2ad4e1SGerd Hoffmann } 1607d2ad4e1SGerd Hoffmann g->parent_obj.enable = 1; 1617d2ad4e1SGerd Hoffmann 1627d2ad4e1SGerd Hoffmann memset(&info, 0, sizeof(info)); 1637d2ad4e1SGerd Hoffmann 1647d2ad4e1SGerd Hoffmann if (ss.resource_id && ss.r.width && ss.r.height) { 1657d2ad4e1SGerd Hoffmann ret = virgl_renderer_resource_get_info(ss.resource_id, &info); 1667d2ad4e1SGerd Hoffmann if (ret == -1) { 1677d2ad4e1SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 1687d2ad4e1SGerd Hoffmann "%s: illegal resource specified %d\n", 1697d2ad4e1SGerd Hoffmann __func__, ss.resource_id); 1707d2ad4e1SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 1717d2ad4e1SGerd Hoffmann return; 1727d2ad4e1SGerd Hoffmann } 1737d2ad4e1SGerd Hoffmann qemu_console_resize(g->parent_obj.scanout[ss.scanout_id].con, 1747d2ad4e1SGerd Hoffmann ss.r.width, ss.r.height); 1757d2ad4e1SGerd Hoffmann virgl_renderer_force_ctx_0(); 1767d2ad4e1SGerd Hoffmann dpy_gl_scanout_texture( 1777d2ad4e1SGerd Hoffmann g->parent_obj.scanout[ss.scanout_id].con, info.tex_id, 178*46e4609eSMarc-André Lureau info.flags & VIRTIO_GPU_RESOURCE_FLAG_Y_0_TOP, 1797d2ad4e1SGerd Hoffmann info.width, info.height, 1807d2ad4e1SGerd Hoffmann ss.r.x, ss.r.y, ss.r.width, ss.r.height); 1817d2ad4e1SGerd Hoffmann } else { 1827d2ad4e1SGerd Hoffmann dpy_gfx_replace_surface( 1837d2ad4e1SGerd Hoffmann g->parent_obj.scanout[ss.scanout_id].con, NULL); 1847d2ad4e1SGerd Hoffmann dpy_gl_scanout_disable(g->parent_obj.scanout[ss.scanout_id].con); 1857d2ad4e1SGerd Hoffmann } 1867d2ad4e1SGerd Hoffmann g->parent_obj.scanout[ss.scanout_id].resource_id = ss.resource_id; 1877d2ad4e1SGerd Hoffmann } 1887d2ad4e1SGerd Hoffmann 1897d2ad4e1SGerd Hoffmann static void virgl_cmd_submit_3d(VirtIOGPU *g, 1907d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 1917d2ad4e1SGerd Hoffmann { 1927d2ad4e1SGerd Hoffmann struct virtio_gpu_cmd_submit cs; 1937d2ad4e1SGerd Hoffmann void *buf; 1947d2ad4e1SGerd Hoffmann size_t s; 1957d2ad4e1SGerd Hoffmann 1967d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(cs); 1977d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_ctx_submit(cs.hdr.ctx_id, cs.size); 1987d2ad4e1SGerd Hoffmann 1997d2ad4e1SGerd Hoffmann buf = g_malloc(cs.size); 2007d2ad4e1SGerd Hoffmann s = iov_to_buf(cmd->elem.out_sg, cmd->elem.out_num, 2017d2ad4e1SGerd Hoffmann sizeof(cs), buf, cs.size); 2027d2ad4e1SGerd Hoffmann if (s != cs.size) { 2037d2ad4e1SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: size mismatch (%zd/%d)", 2047d2ad4e1SGerd Hoffmann __func__, s, cs.size); 2057d2ad4e1SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 2067d2ad4e1SGerd Hoffmann goto out; 2077d2ad4e1SGerd Hoffmann } 2087d2ad4e1SGerd Hoffmann 2097d2ad4e1SGerd Hoffmann if (virtio_gpu_stats_enabled(g->parent_obj.conf)) { 2107d2ad4e1SGerd Hoffmann g->stats.req_3d++; 2117d2ad4e1SGerd Hoffmann g->stats.bytes_3d += cs.size; 2127d2ad4e1SGerd Hoffmann } 2137d2ad4e1SGerd Hoffmann 2147d2ad4e1SGerd Hoffmann virgl_renderer_submit_cmd(buf, cs.hdr.ctx_id, cs.size / 4); 2157d2ad4e1SGerd Hoffmann 2167d2ad4e1SGerd Hoffmann out: 2177d2ad4e1SGerd Hoffmann g_free(buf); 2187d2ad4e1SGerd Hoffmann } 2197d2ad4e1SGerd Hoffmann 2207d2ad4e1SGerd Hoffmann static void virgl_cmd_transfer_to_host_2d(VirtIOGPU *g, 2217d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 2227d2ad4e1SGerd Hoffmann { 2237d2ad4e1SGerd Hoffmann struct virtio_gpu_transfer_to_host_2d t2d; 2247d2ad4e1SGerd Hoffmann struct virtio_gpu_box box; 2257d2ad4e1SGerd Hoffmann 2267d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(t2d); 2277d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_xfer_toh_2d(t2d.resource_id); 2287d2ad4e1SGerd Hoffmann 2297d2ad4e1SGerd Hoffmann box.x = t2d.r.x; 2307d2ad4e1SGerd Hoffmann box.y = t2d.r.y; 2317d2ad4e1SGerd Hoffmann box.z = 0; 2327d2ad4e1SGerd Hoffmann box.w = t2d.r.width; 2337d2ad4e1SGerd Hoffmann box.h = t2d.r.height; 2347d2ad4e1SGerd Hoffmann box.d = 1; 2357d2ad4e1SGerd Hoffmann 2367d2ad4e1SGerd Hoffmann virgl_renderer_transfer_write_iov(t2d.resource_id, 2377d2ad4e1SGerd Hoffmann 0, 2387d2ad4e1SGerd Hoffmann 0, 2397d2ad4e1SGerd Hoffmann 0, 2407d2ad4e1SGerd Hoffmann 0, 2417d2ad4e1SGerd Hoffmann (struct virgl_box *)&box, 2427d2ad4e1SGerd Hoffmann t2d.offset, NULL, 0); 2437d2ad4e1SGerd Hoffmann } 2447d2ad4e1SGerd Hoffmann 2457d2ad4e1SGerd Hoffmann static void virgl_cmd_transfer_to_host_3d(VirtIOGPU *g, 2467d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 2477d2ad4e1SGerd Hoffmann { 2487d2ad4e1SGerd Hoffmann struct virtio_gpu_transfer_host_3d t3d; 2497d2ad4e1SGerd Hoffmann 2507d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(t3d); 2517d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_xfer_toh_3d(t3d.resource_id); 2527d2ad4e1SGerd Hoffmann 2537d2ad4e1SGerd Hoffmann virgl_renderer_transfer_write_iov(t3d.resource_id, 2547d2ad4e1SGerd Hoffmann t3d.hdr.ctx_id, 2557d2ad4e1SGerd Hoffmann t3d.level, 2567d2ad4e1SGerd Hoffmann t3d.stride, 2577d2ad4e1SGerd Hoffmann t3d.layer_stride, 2587d2ad4e1SGerd Hoffmann (struct virgl_box *)&t3d.box, 2597d2ad4e1SGerd Hoffmann t3d.offset, NULL, 0); 2607d2ad4e1SGerd Hoffmann } 2617d2ad4e1SGerd Hoffmann 2627d2ad4e1SGerd Hoffmann static void 2637d2ad4e1SGerd Hoffmann virgl_cmd_transfer_from_host_3d(VirtIOGPU *g, 2647d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 2657d2ad4e1SGerd Hoffmann { 2667d2ad4e1SGerd Hoffmann struct virtio_gpu_transfer_host_3d tf3d; 2677d2ad4e1SGerd Hoffmann 2687d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(tf3d); 2697d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_xfer_fromh_3d(tf3d.resource_id); 2707d2ad4e1SGerd Hoffmann 2717d2ad4e1SGerd Hoffmann virgl_renderer_transfer_read_iov(tf3d.resource_id, 2727d2ad4e1SGerd Hoffmann tf3d.hdr.ctx_id, 2737d2ad4e1SGerd Hoffmann tf3d.level, 2747d2ad4e1SGerd Hoffmann tf3d.stride, 2757d2ad4e1SGerd Hoffmann tf3d.layer_stride, 2767d2ad4e1SGerd Hoffmann (struct virgl_box *)&tf3d.box, 2777d2ad4e1SGerd Hoffmann tf3d.offset, NULL, 0); 2787d2ad4e1SGerd Hoffmann } 2797d2ad4e1SGerd Hoffmann 2807d2ad4e1SGerd Hoffmann 2817d2ad4e1SGerd Hoffmann static void virgl_resource_attach_backing(VirtIOGPU *g, 2827d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 2837d2ad4e1SGerd Hoffmann { 2847d2ad4e1SGerd Hoffmann struct virtio_gpu_resource_attach_backing att_rb; 2857d2ad4e1SGerd Hoffmann struct iovec *res_iovs; 2867d2ad4e1SGerd Hoffmann uint32_t res_niov; 2877d2ad4e1SGerd Hoffmann int ret; 2887d2ad4e1SGerd Hoffmann 2897d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(att_rb); 2907d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_back_attach(att_rb.resource_id); 2917d2ad4e1SGerd Hoffmann 29270d37662SVivek Kasireddy ret = virtio_gpu_create_mapping_iov(g, att_rb.nr_entries, sizeof(att_rb), 29370d37662SVivek Kasireddy cmd, NULL, &res_iovs, &res_niov); 2947d2ad4e1SGerd Hoffmann if (ret != 0) { 2957d2ad4e1SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 2967d2ad4e1SGerd Hoffmann return; 2977d2ad4e1SGerd Hoffmann } 2987d2ad4e1SGerd Hoffmann 2997d2ad4e1SGerd Hoffmann ret = virgl_renderer_resource_attach_iov(att_rb.resource_id, 3007d2ad4e1SGerd Hoffmann res_iovs, res_niov); 3017d2ad4e1SGerd Hoffmann 3027d2ad4e1SGerd Hoffmann if (ret != 0) 3037d2ad4e1SGerd Hoffmann virtio_gpu_cleanup_mapping_iov(g, res_iovs, res_niov); 3047d2ad4e1SGerd Hoffmann } 3057d2ad4e1SGerd Hoffmann 3067d2ad4e1SGerd Hoffmann static void virgl_resource_detach_backing(VirtIOGPU *g, 3077d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 3087d2ad4e1SGerd Hoffmann { 3097d2ad4e1SGerd Hoffmann struct virtio_gpu_resource_detach_backing detach_rb; 3107d2ad4e1SGerd Hoffmann struct iovec *res_iovs = NULL; 3117d2ad4e1SGerd Hoffmann int num_iovs = 0; 3127d2ad4e1SGerd Hoffmann 3137d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(detach_rb); 3147d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_res_back_detach(detach_rb.resource_id); 3157d2ad4e1SGerd Hoffmann 3167d2ad4e1SGerd Hoffmann virgl_renderer_resource_detach_iov(detach_rb.resource_id, 3177d2ad4e1SGerd Hoffmann &res_iovs, 3187d2ad4e1SGerd Hoffmann &num_iovs); 3197d2ad4e1SGerd Hoffmann if (res_iovs == NULL || num_iovs == 0) { 3207d2ad4e1SGerd Hoffmann return; 3217d2ad4e1SGerd Hoffmann } 3227d2ad4e1SGerd Hoffmann virtio_gpu_cleanup_mapping_iov(g, res_iovs, num_iovs); 3237d2ad4e1SGerd Hoffmann } 3247d2ad4e1SGerd Hoffmann 3257d2ad4e1SGerd Hoffmann 3267d2ad4e1SGerd Hoffmann static void virgl_cmd_ctx_attach_resource(VirtIOGPU *g, 3277d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 3287d2ad4e1SGerd Hoffmann { 3297d2ad4e1SGerd Hoffmann struct virtio_gpu_ctx_resource att_res; 3307d2ad4e1SGerd Hoffmann 3317d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(att_res); 3327d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_ctx_res_attach(att_res.hdr.ctx_id, 3337d2ad4e1SGerd Hoffmann att_res.resource_id); 3347d2ad4e1SGerd Hoffmann 3357d2ad4e1SGerd Hoffmann virgl_renderer_ctx_attach_resource(att_res.hdr.ctx_id, att_res.resource_id); 3367d2ad4e1SGerd Hoffmann } 3377d2ad4e1SGerd Hoffmann 3387d2ad4e1SGerd Hoffmann static void virgl_cmd_ctx_detach_resource(VirtIOGPU *g, 3397d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 3407d2ad4e1SGerd Hoffmann { 3417d2ad4e1SGerd Hoffmann struct virtio_gpu_ctx_resource det_res; 3427d2ad4e1SGerd Hoffmann 3437d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(det_res); 3447d2ad4e1SGerd Hoffmann trace_virtio_gpu_cmd_ctx_res_detach(det_res.hdr.ctx_id, 3457d2ad4e1SGerd Hoffmann det_res.resource_id); 3467d2ad4e1SGerd Hoffmann 3477d2ad4e1SGerd Hoffmann virgl_renderer_ctx_detach_resource(det_res.hdr.ctx_id, det_res.resource_id); 3487d2ad4e1SGerd Hoffmann } 3497d2ad4e1SGerd Hoffmann 3507d2ad4e1SGerd Hoffmann static void virgl_cmd_get_capset_info(VirtIOGPU *g, 3517d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 3527d2ad4e1SGerd Hoffmann { 3537d2ad4e1SGerd Hoffmann struct virtio_gpu_get_capset_info info; 3547d2ad4e1SGerd Hoffmann struct virtio_gpu_resp_capset_info resp; 3557d2ad4e1SGerd Hoffmann 3567d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(info); 3577d2ad4e1SGerd Hoffmann 3587d2ad4e1SGerd Hoffmann memset(&resp, 0, sizeof(resp)); 3597d2ad4e1SGerd Hoffmann if (info.capset_index == 0) { 3607d2ad4e1SGerd Hoffmann resp.capset_id = VIRTIO_GPU_CAPSET_VIRGL; 3617d2ad4e1SGerd Hoffmann virgl_renderer_get_cap_set(resp.capset_id, 3627d2ad4e1SGerd Hoffmann &resp.capset_max_version, 3637d2ad4e1SGerd Hoffmann &resp.capset_max_size); 3647d2ad4e1SGerd Hoffmann } else if (info.capset_index == 1) { 3657d2ad4e1SGerd Hoffmann resp.capset_id = VIRTIO_GPU_CAPSET_VIRGL2; 3667d2ad4e1SGerd Hoffmann virgl_renderer_get_cap_set(resp.capset_id, 3677d2ad4e1SGerd Hoffmann &resp.capset_max_version, 3687d2ad4e1SGerd Hoffmann &resp.capset_max_size); 3697d2ad4e1SGerd Hoffmann } else { 3707d2ad4e1SGerd Hoffmann resp.capset_max_version = 0; 3717d2ad4e1SGerd Hoffmann resp.capset_max_size = 0; 3727d2ad4e1SGerd Hoffmann } 3737d2ad4e1SGerd Hoffmann resp.hdr.type = VIRTIO_GPU_RESP_OK_CAPSET_INFO; 3747d2ad4e1SGerd Hoffmann virtio_gpu_ctrl_response(g, cmd, &resp.hdr, sizeof(resp)); 3757d2ad4e1SGerd Hoffmann } 3767d2ad4e1SGerd Hoffmann 3777d2ad4e1SGerd Hoffmann static void virgl_cmd_get_capset(VirtIOGPU *g, 3787d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 3797d2ad4e1SGerd Hoffmann { 3807d2ad4e1SGerd Hoffmann struct virtio_gpu_get_capset gc; 3817d2ad4e1SGerd Hoffmann struct virtio_gpu_resp_capset *resp; 3827d2ad4e1SGerd Hoffmann uint32_t max_ver, max_size; 3837d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(gc); 3847d2ad4e1SGerd Hoffmann 3857d2ad4e1SGerd Hoffmann virgl_renderer_get_cap_set(gc.capset_id, &max_ver, 3867d2ad4e1SGerd Hoffmann &max_size); 3877d2ad4e1SGerd Hoffmann if (!max_size) { 3887d2ad4e1SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 3897d2ad4e1SGerd Hoffmann return; 3907d2ad4e1SGerd Hoffmann } 3917d2ad4e1SGerd Hoffmann 3927d2ad4e1SGerd Hoffmann resp = g_malloc0(sizeof(*resp) + max_size); 3937d2ad4e1SGerd Hoffmann resp->hdr.type = VIRTIO_GPU_RESP_OK_CAPSET; 3947d2ad4e1SGerd Hoffmann virgl_renderer_fill_caps(gc.capset_id, 3957d2ad4e1SGerd Hoffmann gc.capset_version, 3967d2ad4e1SGerd Hoffmann (void *)resp->capset_data); 3977d2ad4e1SGerd Hoffmann virtio_gpu_ctrl_response(g, cmd, &resp->hdr, sizeof(*resp) + max_size); 3987d2ad4e1SGerd Hoffmann g_free(resp); 3997d2ad4e1SGerd Hoffmann } 4007d2ad4e1SGerd Hoffmann 4017d2ad4e1SGerd Hoffmann void virtio_gpu_virgl_process_cmd(VirtIOGPU *g, 4027d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 4037d2ad4e1SGerd Hoffmann { 4047d2ad4e1SGerd Hoffmann VIRTIO_GPU_FILL_CMD(cmd->cmd_hdr); 4057d2ad4e1SGerd Hoffmann 4067d2ad4e1SGerd Hoffmann virgl_renderer_force_ctx_0(); 4077d2ad4e1SGerd Hoffmann switch (cmd->cmd_hdr.type) { 4087d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_CTX_CREATE: 4097d2ad4e1SGerd Hoffmann virgl_cmd_context_create(g, cmd); 4107d2ad4e1SGerd Hoffmann break; 4117d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_CTX_DESTROY: 4127d2ad4e1SGerd Hoffmann virgl_cmd_context_destroy(g, cmd); 4137d2ad4e1SGerd Hoffmann break; 4147d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_CREATE_2D: 4157d2ad4e1SGerd Hoffmann virgl_cmd_create_resource_2d(g, cmd); 4167d2ad4e1SGerd Hoffmann break; 4177d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_CREATE_3D: 4187d2ad4e1SGerd Hoffmann virgl_cmd_create_resource_3d(g, cmd); 4197d2ad4e1SGerd Hoffmann break; 4207d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_SUBMIT_3D: 4217d2ad4e1SGerd Hoffmann virgl_cmd_submit_3d(g, cmd); 4227d2ad4e1SGerd Hoffmann break; 4237d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_TRANSFER_TO_HOST_2D: 4247d2ad4e1SGerd Hoffmann virgl_cmd_transfer_to_host_2d(g, cmd); 4257d2ad4e1SGerd Hoffmann break; 4267d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_TRANSFER_TO_HOST_3D: 4277d2ad4e1SGerd Hoffmann virgl_cmd_transfer_to_host_3d(g, cmd); 4287d2ad4e1SGerd Hoffmann break; 4297d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_TRANSFER_FROM_HOST_3D: 4307d2ad4e1SGerd Hoffmann virgl_cmd_transfer_from_host_3d(g, cmd); 4317d2ad4e1SGerd Hoffmann break; 4327d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_ATTACH_BACKING: 4337d2ad4e1SGerd Hoffmann virgl_resource_attach_backing(g, cmd); 4347d2ad4e1SGerd Hoffmann break; 4357d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_DETACH_BACKING: 4367d2ad4e1SGerd Hoffmann virgl_resource_detach_backing(g, cmd); 4377d2ad4e1SGerd Hoffmann break; 4387d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_SET_SCANOUT: 4397d2ad4e1SGerd Hoffmann virgl_cmd_set_scanout(g, cmd); 4407d2ad4e1SGerd Hoffmann break; 4417d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_FLUSH: 4427d2ad4e1SGerd Hoffmann virgl_cmd_resource_flush(g, cmd); 4437d2ad4e1SGerd Hoffmann break; 4447d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_UNREF: 4457d2ad4e1SGerd Hoffmann virgl_cmd_resource_unref(g, cmd); 4467d2ad4e1SGerd Hoffmann break; 4477d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_CTX_ATTACH_RESOURCE: 4487d2ad4e1SGerd Hoffmann /* TODO add security */ 4497d2ad4e1SGerd Hoffmann virgl_cmd_ctx_attach_resource(g, cmd); 4507d2ad4e1SGerd Hoffmann break; 4517d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_CTX_DETACH_RESOURCE: 4527d2ad4e1SGerd Hoffmann /* TODO add security */ 4537d2ad4e1SGerd Hoffmann virgl_cmd_ctx_detach_resource(g, cmd); 4547d2ad4e1SGerd Hoffmann break; 4557d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_GET_CAPSET_INFO: 4567d2ad4e1SGerd Hoffmann virgl_cmd_get_capset_info(g, cmd); 4577d2ad4e1SGerd Hoffmann break; 4587d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_GET_CAPSET: 4597d2ad4e1SGerd Hoffmann virgl_cmd_get_capset(g, cmd); 4607d2ad4e1SGerd Hoffmann break; 4617d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_GET_DISPLAY_INFO: 4627d2ad4e1SGerd Hoffmann virtio_gpu_get_display_info(g, cmd); 4637d2ad4e1SGerd Hoffmann break; 4647d2ad4e1SGerd Hoffmann case VIRTIO_GPU_CMD_GET_EDID: 4657d2ad4e1SGerd Hoffmann virtio_gpu_get_edid(g, cmd); 4667d2ad4e1SGerd Hoffmann break; 4677d2ad4e1SGerd Hoffmann default: 4687d2ad4e1SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 4697d2ad4e1SGerd Hoffmann break; 4707d2ad4e1SGerd Hoffmann } 4717d2ad4e1SGerd Hoffmann 4727d2ad4e1SGerd Hoffmann if (cmd->finished) { 4737d2ad4e1SGerd Hoffmann return; 4747d2ad4e1SGerd Hoffmann } 4757d2ad4e1SGerd Hoffmann if (cmd->error) { 4767d2ad4e1SGerd Hoffmann fprintf(stderr, "%s: ctrl 0x%x, error 0x%x\n", __func__, 4777d2ad4e1SGerd Hoffmann cmd->cmd_hdr.type, cmd->error); 4787d2ad4e1SGerd Hoffmann virtio_gpu_ctrl_response_nodata(g, cmd, cmd->error); 4797d2ad4e1SGerd Hoffmann return; 4807d2ad4e1SGerd Hoffmann } 4817d2ad4e1SGerd Hoffmann if (!(cmd->cmd_hdr.flags & VIRTIO_GPU_FLAG_FENCE)) { 4827d2ad4e1SGerd Hoffmann virtio_gpu_ctrl_response_nodata(g, cmd, VIRTIO_GPU_RESP_OK_NODATA); 4837d2ad4e1SGerd Hoffmann return; 4847d2ad4e1SGerd Hoffmann } 4857d2ad4e1SGerd Hoffmann 4867d2ad4e1SGerd Hoffmann trace_virtio_gpu_fence_ctrl(cmd->cmd_hdr.fence_id, cmd->cmd_hdr.type); 4877d2ad4e1SGerd Hoffmann virgl_renderer_create_fence(cmd->cmd_hdr.fence_id, cmd->cmd_hdr.type); 4887d2ad4e1SGerd Hoffmann } 4897d2ad4e1SGerd Hoffmann 4907d2ad4e1SGerd Hoffmann static void virgl_write_fence(void *opaque, uint32_t fence) 4917d2ad4e1SGerd Hoffmann { 4927d2ad4e1SGerd Hoffmann VirtIOGPU *g = opaque; 4937d2ad4e1SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd, *tmp; 4947d2ad4e1SGerd Hoffmann 4957d2ad4e1SGerd Hoffmann QTAILQ_FOREACH_SAFE(cmd, &g->fenceq, next, tmp) { 4967d2ad4e1SGerd Hoffmann /* 4977d2ad4e1SGerd Hoffmann * the guest can end up emitting fences out of order 4987d2ad4e1SGerd Hoffmann * so we should check all fenced cmds not just the first one. 4997d2ad4e1SGerd Hoffmann */ 5007d2ad4e1SGerd Hoffmann if (cmd->cmd_hdr.fence_id > fence) { 5017d2ad4e1SGerd Hoffmann continue; 5027d2ad4e1SGerd Hoffmann } 5037d2ad4e1SGerd Hoffmann trace_virtio_gpu_fence_resp(cmd->cmd_hdr.fence_id); 5047d2ad4e1SGerd Hoffmann virtio_gpu_ctrl_response_nodata(g, cmd, VIRTIO_GPU_RESP_OK_NODATA); 5057d2ad4e1SGerd Hoffmann QTAILQ_REMOVE(&g->fenceq, cmd, next); 5067d2ad4e1SGerd Hoffmann g_free(cmd); 5077d2ad4e1SGerd Hoffmann g->inflight--; 5087d2ad4e1SGerd Hoffmann if (virtio_gpu_stats_enabled(g->parent_obj.conf)) { 5097d2ad4e1SGerd Hoffmann fprintf(stderr, "inflight: %3d (-)\r", g->inflight); 5107d2ad4e1SGerd Hoffmann } 5117d2ad4e1SGerd Hoffmann } 5127d2ad4e1SGerd Hoffmann } 5137d2ad4e1SGerd Hoffmann 5147d2ad4e1SGerd Hoffmann static virgl_renderer_gl_context 5157d2ad4e1SGerd Hoffmann virgl_create_context(void *opaque, int scanout_idx, 5167d2ad4e1SGerd Hoffmann struct virgl_renderer_gl_ctx_param *params) 5177d2ad4e1SGerd Hoffmann { 5187d2ad4e1SGerd Hoffmann VirtIOGPU *g = opaque; 5197d2ad4e1SGerd Hoffmann QEMUGLContext ctx; 5207d2ad4e1SGerd Hoffmann QEMUGLParams qparams; 5217d2ad4e1SGerd Hoffmann 5227d2ad4e1SGerd Hoffmann qparams.major_ver = params->major_ver; 5237d2ad4e1SGerd Hoffmann qparams.minor_ver = params->minor_ver; 5247d2ad4e1SGerd Hoffmann 5257d2ad4e1SGerd Hoffmann ctx = dpy_gl_ctx_create(g->parent_obj.scanout[scanout_idx].con, &qparams); 5267d2ad4e1SGerd Hoffmann return (virgl_renderer_gl_context)ctx; 5277d2ad4e1SGerd Hoffmann } 5287d2ad4e1SGerd Hoffmann 5297d2ad4e1SGerd Hoffmann static void virgl_destroy_context(void *opaque, virgl_renderer_gl_context ctx) 5307d2ad4e1SGerd Hoffmann { 5317d2ad4e1SGerd Hoffmann VirtIOGPU *g = opaque; 5327d2ad4e1SGerd Hoffmann QEMUGLContext qctx = (QEMUGLContext)ctx; 5337d2ad4e1SGerd Hoffmann 5347d2ad4e1SGerd Hoffmann dpy_gl_ctx_destroy(g->parent_obj.scanout[0].con, qctx); 5357d2ad4e1SGerd Hoffmann } 5367d2ad4e1SGerd Hoffmann 5377d2ad4e1SGerd Hoffmann static int virgl_make_context_current(void *opaque, int scanout_idx, 5387d2ad4e1SGerd Hoffmann virgl_renderer_gl_context ctx) 5397d2ad4e1SGerd Hoffmann { 5407d2ad4e1SGerd Hoffmann VirtIOGPU *g = opaque; 5417d2ad4e1SGerd Hoffmann QEMUGLContext qctx = (QEMUGLContext)ctx; 5427d2ad4e1SGerd Hoffmann 5437d2ad4e1SGerd Hoffmann return dpy_gl_ctx_make_current(g->parent_obj.scanout[scanout_idx].con, 5447d2ad4e1SGerd Hoffmann qctx); 5457d2ad4e1SGerd Hoffmann } 5467d2ad4e1SGerd Hoffmann 5477d2ad4e1SGerd Hoffmann static struct virgl_renderer_callbacks virtio_gpu_3d_cbs = { 5487d2ad4e1SGerd Hoffmann .version = 1, 5497d2ad4e1SGerd Hoffmann .write_fence = virgl_write_fence, 5507d2ad4e1SGerd Hoffmann .create_gl_context = virgl_create_context, 5517d2ad4e1SGerd Hoffmann .destroy_gl_context = virgl_destroy_context, 5527d2ad4e1SGerd Hoffmann .make_current = virgl_make_context_current, 5537d2ad4e1SGerd Hoffmann }; 5547d2ad4e1SGerd Hoffmann 5557d2ad4e1SGerd Hoffmann static void virtio_gpu_print_stats(void *opaque) 5567d2ad4e1SGerd Hoffmann { 5577d2ad4e1SGerd Hoffmann VirtIOGPU *g = opaque; 5587d2ad4e1SGerd Hoffmann 5597d2ad4e1SGerd Hoffmann if (g->stats.requests) { 5607d2ad4e1SGerd Hoffmann fprintf(stderr, "stats: vq req %4d, %3d -- 3D %4d (%5d)\n", 5617d2ad4e1SGerd Hoffmann g->stats.requests, 5627d2ad4e1SGerd Hoffmann g->stats.max_inflight, 5637d2ad4e1SGerd Hoffmann g->stats.req_3d, 5647d2ad4e1SGerd Hoffmann g->stats.bytes_3d); 5657d2ad4e1SGerd Hoffmann g->stats.requests = 0; 5667d2ad4e1SGerd Hoffmann g->stats.max_inflight = 0; 5677d2ad4e1SGerd Hoffmann g->stats.req_3d = 0; 5687d2ad4e1SGerd Hoffmann g->stats.bytes_3d = 0; 5697d2ad4e1SGerd Hoffmann } else { 5707d2ad4e1SGerd Hoffmann fprintf(stderr, "stats: idle\r"); 5717d2ad4e1SGerd Hoffmann } 5727d2ad4e1SGerd Hoffmann timer_mod(g->print_stats, qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL) + 1000); 5737d2ad4e1SGerd Hoffmann } 5747d2ad4e1SGerd Hoffmann 5757d2ad4e1SGerd Hoffmann static void virtio_gpu_fence_poll(void *opaque) 5767d2ad4e1SGerd Hoffmann { 5777d2ad4e1SGerd Hoffmann VirtIOGPU *g = opaque; 5787d2ad4e1SGerd Hoffmann 5797d2ad4e1SGerd Hoffmann virgl_renderer_poll(); 5807d2ad4e1SGerd Hoffmann virtio_gpu_process_cmdq(g); 5817d2ad4e1SGerd Hoffmann if (!QTAILQ_EMPTY(&g->cmdq) || !QTAILQ_EMPTY(&g->fenceq)) { 5827d2ad4e1SGerd Hoffmann timer_mod(g->fence_poll, qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL) + 10); 5837d2ad4e1SGerd Hoffmann } 5847d2ad4e1SGerd Hoffmann } 5857d2ad4e1SGerd Hoffmann 5867d2ad4e1SGerd Hoffmann void virtio_gpu_virgl_fence_poll(VirtIOGPU *g) 5877d2ad4e1SGerd Hoffmann { 5887d2ad4e1SGerd Hoffmann virtio_gpu_fence_poll(g); 5897d2ad4e1SGerd Hoffmann } 5907d2ad4e1SGerd Hoffmann 5918a13b9bcSMarc-André Lureau void virtio_gpu_virgl_reset_scanout(VirtIOGPU *g) 5927d2ad4e1SGerd Hoffmann { 5937d2ad4e1SGerd Hoffmann int i; 5947d2ad4e1SGerd Hoffmann 5957d2ad4e1SGerd Hoffmann for (i = 0; i < g->parent_obj.conf.max_outputs; i++) { 5967d2ad4e1SGerd Hoffmann dpy_gfx_replace_surface(g->parent_obj.scanout[i].con, NULL); 5977d2ad4e1SGerd Hoffmann dpy_gl_scanout_disable(g->parent_obj.scanout[i].con); 5987d2ad4e1SGerd Hoffmann } 5997d2ad4e1SGerd Hoffmann } 6007d2ad4e1SGerd Hoffmann 6018a13b9bcSMarc-André Lureau void virtio_gpu_virgl_reset(VirtIOGPU *g) 6028a13b9bcSMarc-André Lureau { 6038a13b9bcSMarc-André Lureau virgl_renderer_reset(); 6048a13b9bcSMarc-André Lureau } 6058a13b9bcSMarc-André Lureau 6067d2ad4e1SGerd Hoffmann int virtio_gpu_virgl_init(VirtIOGPU *g) 6077d2ad4e1SGerd Hoffmann { 6087d2ad4e1SGerd Hoffmann int ret; 6097d2ad4e1SGerd Hoffmann 6107d2ad4e1SGerd Hoffmann ret = virgl_renderer_init(g, 0, &virtio_gpu_3d_cbs); 6117d2ad4e1SGerd Hoffmann if (ret != 0) { 6128f5f1ea0SMarc-André Lureau error_report("virgl could not be initialized: %d", ret); 6137d2ad4e1SGerd Hoffmann return ret; 6147d2ad4e1SGerd Hoffmann } 6157d2ad4e1SGerd Hoffmann 6167d2ad4e1SGerd Hoffmann g->fence_poll = timer_new_ms(QEMU_CLOCK_VIRTUAL, 6177d2ad4e1SGerd Hoffmann virtio_gpu_fence_poll, g); 6187d2ad4e1SGerd Hoffmann 6197d2ad4e1SGerd Hoffmann if (virtio_gpu_stats_enabled(g->parent_obj.conf)) { 6207d2ad4e1SGerd Hoffmann g->print_stats = timer_new_ms(QEMU_CLOCK_VIRTUAL, 6217d2ad4e1SGerd Hoffmann virtio_gpu_print_stats, g); 6227d2ad4e1SGerd Hoffmann timer_mod(g->print_stats, qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL) + 1000); 6237d2ad4e1SGerd Hoffmann } 6247d2ad4e1SGerd Hoffmann return 0; 6257d2ad4e1SGerd Hoffmann } 6267d2ad4e1SGerd Hoffmann 6277d2ad4e1SGerd Hoffmann int virtio_gpu_virgl_get_num_capsets(VirtIOGPU *g) 6287d2ad4e1SGerd Hoffmann { 6297d2ad4e1SGerd Hoffmann uint32_t capset2_max_ver, capset2_max_size; 6307d2ad4e1SGerd Hoffmann virgl_renderer_get_cap_set(VIRTIO_GPU_CAPSET_VIRGL2, 6317d2ad4e1SGerd Hoffmann &capset2_max_ver, 6327d2ad4e1SGerd Hoffmann &capset2_max_size); 6337d2ad4e1SGerd Hoffmann 6347d2ad4e1SGerd Hoffmann return capset2_max_ver ? 2 : 1; 6357d2ad4e1SGerd Hoffmann } 636