162232bf4SGerd Hoffmann /* 262232bf4SGerd Hoffmann * Virtio GPU Device 362232bf4SGerd Hoffmann * 462232bf4SGerd Hoffmann * Copyright Red Hat, Inc. 2013-2014 562232bf4SGerd Hoffmann * 662232bf4SGerd Hoffmann * Authors: 762232bf4SGerd Hoffmann * Dave Airlie <airlied@redhat.com> 862232bf4SGerd Hoffmann * Gerd Hoffmann <kraxel@redhat.com> 962232bf4SGerd Hoffmann * 102e252145SGerd Hoffmann * This work is licensed under the terms of the GNU GPL, version 2 or later. 1162232bf4SGerd Hoffmann * See the COPYING file in the top-level directory. 1262232bf4SGerd Hoffmann */ 1362232bf4SGerd Hoffmann 149b8bfe21SPeter Maydell #include "qemu/osdep.h" 1562232bf4SGerd Hoffmann #include "qemu-common.h" 1662232bf4SGerd Hoffmann #include "qemu/iov.h" 1762232bf4SGerd Hoffmann #include "ui/console.h" 1862232bf4SGerd Hoffmann #include "trace.h" 1962232bf4SGerd Hoffmann #include "hw/virtio/virtio.h" 2062232bf4SGerd Hoffmann #include "hw/virtio/virtio-gpu.h" 2162232bf4SGerd Hoffmann #include "hw/virtio/virtio-bus.h" 2203dd024fSPaolo Bonzini #include "qemu/log.h" 235e3d741cSMarc-André Lureau #include "qapi/error.h" 2462232bf4SGerd Hoffmann 25*0c244e50SGerd Hoffmann #define VIRTIO_GPU_VM_VERSION 1 26*0c244e50SGerd Hoffmann 2762232bf4SGerd Hoffmann static struct virtio_gpu_simple_resource* 2862232bf4SGerd Hoffmann virtio_gpu_find_resource(VirtIOGPU *g, uint32_t resource_id); 2962232bf4SGerd Hoffmann 309d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 319d9e1521SGerd Hoffmann #include "virglrenderer.h" 329d9e1521SGerd Hoffmann #define VIRGL(_g, _virgl, _simple, ...) \ 339d9e1521SGerd Hoffmann do { \ 349d9e1521SGerd Hoffmann if (_g->use_virgl_renderer) { \ 359d9e1521SGerd Hoffmann _virgl(__VA_ARGS__); \ 369d9e1521SGerd Hoffmann } else { \ 379d9e1521SGerd Hoffmann _simple(__VA_ARGS__); \ 389d9e1521SGerd Hoffmann } \ 399d9e1521SGerd Hoffmann } while (0) 409d9e1521SGerd Hoffmann #else 419d9e1521SGerd Hoffmann #define VIRGL(_g, _virgl, _simple, ...) \ 429d9e1521SGerd Hoffmann do { \ 439d9e1521SGerd Hoffmann _simple(__VA_ARGS__); \ 449d9e1521SGerd Hoffmann } while (0) 459d9e1521SGerd Hoffmann #endif 469d9e1521SGerd Hoffmann 4762232bf4SGerd Hoffmann static void update_cursor_data_simple(VirtIOGPU *g, 4862232bf4SGerd Hoffmann struct virtio_gpu_scanout *s, 4962232bf4SGerd Hoffmann uint32_t resource_id) 5062232bf4SGerd Hoffmann { 5162232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 5262232bf4SGerd Hoffmann uint32_t pixels; 5362232bf4SGerd Hoffmann 5462232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, resource_id); 5562232bf4SGerd Hoffmann if (!res) { 5662232bf4SGerd Hoffmann return; 5762232bf4SGerd Hoffmann } 5862232bf4SGerd Hoffmann 5962232bf4SGerd Hoffmann if (pixman_image_get_width(res->image) != s->current_cursor->width || 6062232bf4SGerd Hoffmann pixman_image_get_height(res->image) != s->current_cursor->height) { 6162232bf4SGerd Hoffmann return; 6262232bf4SGerd Hoffmann } 6362232bf4SGerd Hoffmann 6462232bf4SGerd Hoffmann pixels = s->current_cursor->width * s->current_cursor->height; 6562232bf4SGerd Hoffmann memcpy(s->current_cursor->data, 6662232bf4SGerd Hoffmann pixman_image_get_data(res->image), 6762232bf4SGerd Hoffmann pixels * sizeof(uint32_t)); 6862232bf4SGerd Hoffmann } 6962232bf4SGerd Hoffmann 709d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 719d9e1521SGerd Hoffmann 729d9e1521SGerd Hoffmann static void update_cursor_data_virgl(VirtIOGPU *g, 739d9e1521SGerd Hoffmann struct virtio_gpu_scanout *s, 749d9e1521SGerd Hoffmann uint32_t resource_id) 759d9e1521SGerd Hoffmann { 769d9e1521SGerd Hoffmann uint32_t width, height; 779d9e1521SGerd Hoffmann uint32_t pixels, *data; 789d9e1521SGerd Hoffmann 799d9e1521SGerd Hoffmann data = virgl_renderer_get_cursor_data(resource_id, &width, &height); 809d9e1521SGerd Hoffmann if (!data) { 819d9e1521SGerd Hoffmann return; 829d9e1521SGerd Hoffmann } 839d9e1521SGerd Hoffmann 849d9e1521SGerd Hoffmann if (width != s->current_cursor->width || 859d9e1521SGerd Hoffmann height != s->current_cursor->height) { 869d9e1521SGerd Hoffmann return; 879d9e1521SGerd Hoffmann } 889d9e1521SGerd Hoffmann 899d9e1521SGerd Hoffmann pixels = s->current_cursor->width * s->current_cursor->height; 909d9e1521SGerd Hoffmann memcpy(s->current_cursor->data, data, pixels * sizeof(uint32_t)); 919d9e1521SGerd Hoffmann free(data); 929d9e1521SGerd Hoffmann } 939d9e1521SGerd Hoffmann 949d9e1521SGerd Hoffmann #endif 959d9e1521SGerd Hoffmann 9662232bf4SGerd Hoffmann static void update_cursor(VirtIOGPU *g, struct virtio_gpu_update_cursor *cursor) 9762232bf4SGerd Hoffmann { 9862232bf4SGerd Hoffmann struct virtio_gpu_scanout *s; 99*0c244e50SGerd Hoffmann bool move = cursor->hdr.type == VIRTIO_GPU_CMD_MOVE_CURSOR; 10062232bf4SGerd Hoffmann 10162232bf4SGerd Hoffmann if (cursor->pos.scanout_id >= g->conf.max_outputs) { 10262232bf4SGerd Hoffmann return; 10362232bf4SGerd Hoffmann } 10462232bf4SGerd Hoffmann s = &g->scanout[cursor->pos.scanout_id]; 10562232bf4SGerd Hoffmann 106e9c1b459SGerd Hoffmann trace_virtio_gpu_update_cursor(cursor->pos.scanout_id, 107e9c1b459SGerd Hoffmann cursor->pos.x, 108e9c1b459SGerd Hoffmann cursor->pos.y, 109e9c1b459SGerd Hoffmann move ? "move" : "update", 110e9c1b459SGerd Hoffmann cursor->resource_id); 111e9c1b459SGerd Hoffmann 112*0c244e50SGerd Hoffmann if (!move) { 11362232bf4SGerd Hoffmann if (!s->current_cursor) { 11462232bf4SGerd Hoffmann s->current_cursor = cursor_alloc(64, 64); 11562232bf4SGerd Hoffmann } 11662232bf4SGerd Hoffmann 11762232bf4SGerd Hoffmann s->current_cursor->hot_x = cursor->hot_x; 11862232bf4SGerd Hoffmann s->current_cursor->hot_y = cursor->hot_y; 11962232bf4SGerd Hoffmann 12062232bf4SGerd Hoffmann if (cursor->resource_id > 0) { 1219d9e1521SGerd Hoffmann VIRGL(g, update_cursor_data_virgl, update_cursor_data_simple, 1229d9e1521SGerd Hoffmann g, s, cursor->resource_id); 12362232bf4SGerd Hoffmann } 12462232bf4SGerd Hoffmann dpy_cursor_define(s->con, s->current_cursor); 125*0c244e50SGerd Hoffmann 126*0c244e50SGerd Hoffmann s->cursor = *cursor; 127*0c244e50SGerd Hoffmann } else { 128*0c244e50SGerd Hoffmann s->cursor.pos.x = cursor->pos.x; 129*0c244e50SGerd Hoffmann s->cursor.pos.y = cursor->pos.y; 13062232bf4SGerd Hoffmann } 13162232bf4SGerd Hoffmann dpy_mouse_set(s->con, cursor->pos.x, cursor->pos.y, 13262232bf4SGerd Hoffmann cursor->resource_id ? 1 : 0); 13362232bf4SGerd Hoffmann } 13462232bf4SGerd Hoffmann 13562232bf4SGerd Hoffmann static void virtio_gpu_get_config(VirtIODevice *vdev, uint8_t *config) 13662232bf4SGerd Hoffmann { 13762232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 13862232bf4SGerd Hoffmann memcpy(config, &g->virtio_config, sizeof(g->virtio_config)); 13962232bf4SGerd Hoffmann } 14062232bf4SGerd Hoffmann 14162232bf4SGerd Hoffmann static void virtio_gpu_set_config(VirtIODevice *vdev, const uint8_t *config) 14262232bf4SGerd Hoffmann { 14362232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 14462232bf4SGerd Hoffmann struct virtio_gpu_config vgconfig; 14562232bf4SGerd Hoffmann 14662232bf4SGerd Hoffmann memcpy(&vgconfig, config, sizeof(g->virtio_config)); 14762232bf4SGerd Hoffmann 14862232bf4SGerd Hoffmann if (vgconfig.events_clear) { 14962232bf4SGerd Hoffmann g->virtio_config.events_read &= ~vgconfig.events_clear; 15062232bf4SGerd Hoffmann } 15162232bf4SGerd Hoffmann } 15262232bf4SGerd Hoffmann 1539d5b731dSJason Wang static uint64_t virtio_gpu_get_features(VirtIODevice *vdev, uint64_t features, 1549d5b731dSJason Wang Error **errp) 15562232bf4SGerd Hoffmann { 1569d9e1521SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 1579d9e1521SGerd Hoffmann 1589d9e1521SGerd Hoffmann if (virtio_gpu_virgl_enabled(g->conf)) { 159fff02bc0SPaolo Bonzini features |= (1 << VIRTIO_GPU_F_VIRGL); 1609d9e1521SGerd Hoffmann } 16162232bf4SGerd Hoffmann return features; 16262232bf4SGerd Hoffmann } 16362232bf4SGerd Hoffmann 1649d9e1521SGerd Hoffmann static void virtio_gpu_set_features(VirtIODevice *vdev, uint64_t features) 1659d9e1521SGerd Hoffmann { 166fff02bc0SPaolo Bonzini static const uint32_t virgl = (1 << VIRTIO_GPU_F_VIRGL); 1679d9e1521SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 1689d9e1521SGerd Hoffmann 1699d9e1521SGerd Hoffmann g->use_virgl_renderer = ((features & virgl) == virgl); 1709d9e1521SGerd Hoffmann trace_virtio_gpu_features(g->use_virgl_renderer); 1719d9e1521SGerd Hoffmann } 1729d9e1521SGerd Hoffmann 17362232bf4SGerd Hoffmann static void virtio_gpu_notify_event(VirtIOGPU *g, uint32_t event_type) 17462232bf4SGerd Hoffmann { 17562232bf4SGerd Hoffmann g->virtio_config.events_read |= event_type; 17662232bf4SGerd Hoffmann virtio_notify_config(&g->parent_obj); 17762232bf4SGerd Hoffmann } 17862232bf4SGerd Hoffmann 17962232bf4SGerd Hoffmann static struct virtio_gpu_simple_resource * 18062232bf4SGerd Hoffmann virtio_gpu_find_resource(VirtIOGPU *g, uint32_t resource_id) 18162232bf4SGerd Hoffmann { 18262232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 18362232bf4SGerd Hoffmann 18462232bf4SGerd Hoffmann QTAILQ_FOREACH(res, &g->reslist, next) { 18562232bf4SGerd Hoffmann if (res->resource_id == resource_id) { 18662232bf4SGerd Hoffmann return res; 18762232bf4SGerd Hoffmann } 18862232bf4SGerd Hoffmann } 18962232bf4SGerd Hoffmann return NULL; 19062232bf4SGerd Hoffmann } 19162232bf4SGerd Hoffmann 19262232bf4SGerd Hoffmann void virtio_gpu_ctrl_response(VirtIOGPU *g, 19362232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd, 19462232bf4SGerd Hoffmann struct virtio_gpu_ctrl_hdr *resp, 19562232bf4SGerd Hoffmann size_t resp_len) 19662232bf4SGerd Hoffmann { 19762232bf4SGerd Hoffmann size_t s; 19862232bf4SGerd Hoffmann 19962232bf4SGerd Hoffmann if (cmd->cmd_hdr.flags & VIRTIO_GPU_FLAG_FENCE) { 20062232bf4SGerd Hoffmann resp->flags |= VIRTIO_GPU_FLAG_FENCE; 20162232bf4SGerd Hoffmann resp->fence_id = cmd->cmd_hdr.fence_id; 20262232bf4SGerd Hoffmann resp->ctx_id = cmd->cmd_hdr.ctx_id; 20362232bf4SGerd Hoffmann } 20462232bf4SGerd Hoffmann s = iov_from_buf(cmd->elem.in_sg, cmd->elem.in_num, 0, resp, resp_len); 20562232bf4SGerd Hoffmann if (s != resp_len) { 20662232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 20762232bf4SGerd Hoffmann "%s: response size incorrect %zu vs %zu\n", 20862232bf4SGerd Hoffmann __func__, s, resp_len); 20962232bf4SGerd Hoffmann } 21062232bf4SGerd Hoffmann virtqueue_push(cmd->vq, &cmd->elem, s); 21162232bf4SGerd Hoffmann virtio_notify(VIRTIO_DEVICE(g), cmd->vq); 21262232bf4SGerd Hoffmann cmd->finished = true; 21362232bf4SGerd Hoffmann } 21462232bf4SGerd Hoffmann 21562232bf4SGerd Hoffmann void virtio_gpu_ctrl_response_nodata(VirtIOGPU *g, 21662232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd, 21762232bf4SGerd Hoffmann enum virtio_gpu_ctrl_type type) 21862232bf4SGerd Hoffmann { 21962232bf4SGerd Hoffmann struct virtio_gpu_ctrl_hdr resp; 22062232bf4SGerd Hoffmann 22162232bf4SGerd Hoffmann memset(&resp, 0, sizeof(resp)); 22262232bf4SGerd Hoffmann resp.type = type; 22362232bf4SGerd Hoffmann virtio_gpu_ctrl_response(g, cmd, &resp, sizeof(resp)); 22462232bf4SGerd Hoffmann } 22562232bf4SGerd Hoffmann 22662232bf4SGerd Hoffmann static void 22762232bf4SGerd Hoffmann virtio_gpu_fill_display_info(VirtIOGPU *g, 22862232bf4SGerd Hoffmann struct virtio_gpu_resp_display_info *dpy_info) 22962232bf4SGerd Hoffmann { 23062232bf4SGerd Hoffmann int i; 23162232bf4SGerd Hoffmann 23262232bf4SGerd Hoffmann for (i = 0; i < g->conf.max_outputs; i++) { 23362232bf4SGerd Hoffmann if (g->enabled_output_bitmask & (1 << i)) { 23462232bf4SGerd Hoffmann dpy_info->pmodes[i].enabled = 1; 23562232bf4SGerd Hoffmann dpy_info->pmodes[i].r.width = g->req_state[i].width; 23662232bf4SGerd Hoffmann dpy_info->pmodes[i].r.height = g->req_state[i].height; 23762232bf4SGerd Hoffmann } 23862232bf4SGerd Hoffmann } 23962232bf4SGerd Hoffmann } 24062232bf4SGerd Hoffmann 24162232bf4SGerd Hoffmann void virtio_gpu_get_display_info(VirtIOGPU *g, 24262232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 24362232bf4SGerd Hoffmann { 24462232bf4SGerd Hoffmann struct virtio_gpu_resp_display_info display_info; 24562232bf4SGerd Hoffmann 24662232bf4SGerd Hoffmann trace_virtio_gpu_cmd_get_display_info(); 24762232bf4SGerd Hoffmann memset(&display_info, 0, sizeof(display_info)); 24862232bf4SGerd Hoffmann display_info.hdr.type = VIRTIO_GPU_RESP_OK_DISPLAY_INFO; 24962232bf4SGerd Hoffmann virtio_gpu_fill_display_info(g, &display_info); 25062232bf4SGerd Hoffmann virtio_gpu_ctrl_response(g, cmd, &display_info.hdr, 25162232bf4SGerd Hoffmann sizeof(display_info)); 25262232bf4SGerd Hoffmann } 25362232bf4SGerd Hoffmann 25462232bf4SGerd Hoffmann static pixman_format_code_t get_pixman_format(uint32_t virtio_gpu_format) 25562232bf4SGerd Hoffmann { 25662232bf4SGerd Hoffmann switch (virtio_gpu_format) { 25762232bf4SGerd Hoffmann #ifdef HOST_WORDS_BIGENDIAN 25862232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8X8_UNORM: 25962232bf4SGerd Hoffmann return PIXMAN_b8g8r8x8; 26062232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8A8_UNORM: 26162232bf4SGerd Hoffmann return PIXMAN_b8g8r8a8; 26262232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8R8G8B8_UNORM: 26362232bf4SGerd Hoffmann return PIXMAN_x8r8g8b8; 26462232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8R8G8B8_UNORM: 26562232bf4SGerd Hoffmann return PIXMAN_a8r8g8b8; 26662232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8X8_UNORM: 26762232bf4SGerd Hoffmann return PIXMAN_r8g8b8x8; 26862232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8A8_UNORM: 26962232bf4SGerd Hoffmann return PIXMAN_r8g8b8a8; 27062232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8B8G8R8_UNORM: 27162232bf4SGerd Hoffmann return PIXMAN_x8b8g8r8; 27262232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8B8G8R8_UNORM: 27362232bf4SGerd Hoffmann return PIXMAN_a8b8g8r8; 27462232bf4SGerd Hoffmann #else 27562232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8X8_UNORM: 27662232bf4SGerd Hoffmann return PIXMAN_x8r8g8b8; 27762232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8A8_UNORM: 27862232bf4SGerd Hoffmann return PIXMAN_a8r8g8b8; 27962232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8R8G8B8_UNORM: 28062232bf4SGerd Hoffmann return PIXMAN_b8g8r8x8; 28162232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8R8G8B8_UNORM: 28262232bf4SGerd Hoffmann return PIXMAN_b8g8r8a8; 28362232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8X8_UNORM: 28462232bf4SGerd Hoffmann return PIXMAN_x8b8g8r8; 28562232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8A8_UNORM: 28662232bf4SGerd Hoffmann return PIXMAN_a8b8g8r8; 28762232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8B8G8R8_UNORM: 28862232bf4SGerd Hoffmann return PIXMAN_r8g8b8x8; 28962232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8B8G8R8_UNORM: 29062232bf4SGerd Hoffmann return PIXMAN_r8g8b8a8; 29162232bf4SGerd Hoffmann #endif 29262232bf4SGerd Hoffmann default: 29362232bf4SGerd Hoffmann return 0; 29462232bf4SGerd Hoffmann } 29562232bf4SGerd Hoffmann } 29662232bf4SGerd Hoffmann 29762232bf4SGerd Hoffmann static void virtio_gpu_resource_create_2d(VirtIOGPU *g, 29862232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 29962232bf4SGerd Hoffmann { 30062232bf4SGerd Hoffmann pixman_format_code_t pformat; 30162232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 30262232bf4SGerd Hoffmann struct virtio_gpu_resource_create_2d c2d; 30362232bf4SGerd Hoffmann 30462232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(c2d); 30562232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_create_2d(c2d.resource_id, c2d.format, 30662232bf4SGerd Hoffmann c2d.width, c2d.height); 30762232bf4SGerd Hoffmann 30862232bf4SGerd Hoffmann if (c2d.resource_id == 0) { 30962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: resource id 0 is not allowed\n", 31062232bf4SGerd Hoffmann __func__); 31162232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 31262232bf4SGerd Hoffmann return; 31362232bf4SGerd Hoffmann } 31462232bf4SGerd Hoffmann 31562232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, c2d.resource_id); 31662232bf4SGerd Hoffmann if (res) { 31762232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: resource already exists %d\n", 31862232bf4SGerd Hoffmann __func__, c2d.resource_id); 31962232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 32062232bf4SGerd Hoffmann return; 32162232bf4SGerd Hoffmann } 32262232bf4SGerd Hoffmann 32362232bf4SGerd Hoffmann res = g_new0(struct virtio_gpu_simple_resource, 1); 32462232bf4SGerd Hoffmann 32562232bf4SGerd Hoffmann res->width = c2d.width; 32662232bf4SGerd Hoffmann res->height = c2d.height; 32762232bf4SGerd Hoffmann res->format = c2d.format; 32862232bf4SGerd Hoffmann res->resource_id = c2d.resource_id; 32962232bf4SGerd Hoffmann 33062232bf4SGerd Hoffmann pformat = get_pixman_format(c2d.format); 33162232bf4SGerd Hoffmann if (!pformat) { 33262232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 33362232bf4SGerd Hoffmann "%s: host couldn't handle guest format %d\n", 33462232bf4SGerd Hoffmann __func__, c2d.format); 33562232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 33662232bf4SGerd Hoffmann return; 33762232bf4SGerd Hoffmann } 33862232bf4SGerd Hoffmann res->image = pixman_image_create_bits(pformat, 33962232bf4SGerd Hoffmann c2d.width, 34062232bf4SGerd Hoffmann c2d.height, 34162232bf4SGerd Hoffmann NULL, 0); 34262232bf4SGerd Hoffmann 34362232bf4SGerd Hoffmann if (!res->image) { 34462232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 34562232bf4SGerd Hoffmann "%s: resource creation failed %d %d %d\n", 34662232bf4SGerd Hoffmann __func__, c2d.resource_id, c2d.width, c2d.height); 34762232bf4SGerd Hoffmann g_free(res); 34862232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_OUT_OF_MEMORY; 34962232bf4SGerd Hoffmann return; 35062232bf4SGerd Hoffmann } 35162232bf4SGerd Hoffmann 35262232bf4SGerd Hoffmann QTAILQ_INSERT_HEAD(&g->reslist, res, next); 35362232bf4SGerd Hoffmann } 35462232bf4SGerd Hoffmann 35562232bf4SGerd Hoffmann static void virtio_gpu_resource_destroy(VirtIOGPU *g, 35662232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res) 35762232bf4SGerd Hoffmann { 35862232bf4SGerd Hoffmann pixman_image_unref(res->image); 35962232bf4SGerd Hoffmann QTAILQ_REMOVE(&g->reslist, res, next); 36062232bf4SGerd Hoffmann g_free(res); 36162232bf4SGerd Hoffmann } 36262232bf4SGerd Hoffmann 36362232bf4SGerd Hoffmann static void virtio_gpu_resource_unref(VirtIOGPU *g, 36462232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 36562232bf4SGerd Hoffmann { 36662232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 36762232bf4SGerd Hoffmann struct virtio_gpu_resource_unref unref; 36862232bf4SGerd Hoffmann 36962232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(unref); 37062232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_unref(unref.resource_id); 37162232bf4SGerd Hoffmann 37262232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, unref.resource_id); 37362232bf4SGerd Hoffmann if (!res) { 37462232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 37562232bf4SGerd Hoffmann __func__, unref.resource_id); 37662232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 37762232bf4SGerd Hoffmann return; 37862232bf4SGerd Hoffmann } 37962232bf4SGerd Hoffmann virtio_gpu_resource_destroy(g, res); 38062232bf4SGerd Hoffmann } 38162232bf4SGerd Hoffmann 38262232bf4SGerd Hoffmann static void virtio_gpu_transfer_to_host_2d(VirtIOGPU *g, 38362232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 38462232bf4SGerd Hoffmann { 38562232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 38662232bf4SGerd Hoffmann int h; 38762232bf4SGerd Hoffmann uint32_t src_offset, dst_offset, stride; 38862232bf4SGerd Hoffmann int bpp; 38962232bf4SGerd Hoffmann pixman_format_code_t format; 39062232bf4SGerd Hoffmann struct virtio_gpu_transfer_to_host_2d t2d; 39162232bf4SGerd Hoffmann 39262232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(t2d); 39362232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_xfer_toh_2d(t2d.resource_id); 39462232bf4SGerd Hoffmann 39562232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, t2d.resource_id); 39662232bf4SGerd Hoffmann if (!res || !res->iov) { 39762232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 39862232bf4SGerd Hoffmann __func__, t2d.resource_id); 39962232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 40062232bf4SGerd Hoffmann return; 40162232bf4SGerd Hoffmann } 40262232bf4SGerd Hoffmann 40362232bf4SGerd Hoffmann if (t2d.r.x > res->width || 40462232bf4SGerd Hoffmann t2d.r.y > res->height || 40562232bf4SGerd Hoffmann t2d.r.width > res->width || 40662232bf4SGerd Hoffmann t2d.r.height > res->height || 40762232bf4SGerd Hoffmann t2d.r.x + t2d.r.width > res->width || 40862232bf4SGerd Hoffmann t2d.r.y + t2d.r.height > res->height) { 40962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: transfer bounds outside resource" 41062232bf4SGerd Hoffmann " bounds for resource %d: %d %d %d %d vs %d %d\n", 41162232bf4SGerd Hoffmann __func__, t2d.resource_id, t2d.r.x, t2d.r.y, 41262232bf4SGerd Hoffmann t2d.r.width, t2d.r.height, res->width, res->height); 41362232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 41462232bf4SGerd Hoffmann return; 41562232bf4SGerd Hoffmann } 41662232bf4SGerd Hoffmann 41762232bf4SGerd Hoffmann format = pixman_image_get_format(res->image); 41862232bf4SGerd Hoffmann bpp = (PIXMAN_FORMAT_BPP(format) + 7) / 8; 41962232bf4SGerd Hoffmann stride = pixman_image_get_stride(res->image); 42062232bf4SGerd Hoffmann 42162232bf4SGerd Hoffmann if (t2d.offset || t2d.r.x || t2d.r.y || 42262232bf4SGerd Hoffmann t2d.r.width != pixman_image_get_width(res->image)) { 42362232bf4SGerd Hoffmann void *img_data = pixman_image_get_data(res->image); 42462232bf4SGerd Hoffmann for (h = 0; h < t2d.r.height; h++) { 42562232bf4SGerd Hoffmann src_offset = t2d.offset + stride * h; 42662232bf4SGerd Hoffmann dst_offset = (t2d.r.y + h) * stride + (t2d.r.x * bpp); 42762232bf4SGerd Hoffmann 42862232bf4SGerd Hoffmann iov_to_buf(res->iov, res->iov_cnt, src_offset, 42962232bf4SGerd Hoffmann (uint8_t *)img_data 43062232bf4SGerd Hoffmann + dst_offset, t2d.r.width * bpp); 43162232bf4SGerd Hoffmann } 43262232bf4SGerd Hoffmann } else { 43362232bf4SGerd Hoffmann iov_to_buf(res->iov, res->iov_cnt, 0, 43462232bf4SGerd Hoffmann pixman_image_get_data(res->image), 43562232bf4SGerd Hoffmann pixman_image_get_stride(res->image) 43662232bf4SGerd Hoffmann * pixman_image_get_height(res->image)); 43762232bf4SGerd Hoffmann } 43862232bf4SGerd Hoffmann } 43962232bf4SGerd Hoffmann 44062232bf4SGerd Hoffmann static void virtio_gpu_resource_flush(VirtIOGPU *g, 44162232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 44262232bf4SGerd Hoffmann { 44362232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 44462232bf4SGerd Hoffmann struct virtio_gpu_resource_flush rf; 44562232bf4SGerd Hoffmann pixman_region16_t flush_region; 44662232bf4SGerd Hoffmann int i; 44762232bf4SGerd Hoffmann 44862232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(rf); 44962232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_flush(rf.resource_id, 45062232bf4SGerd Hoffmann rf.r.width, rf.r.height, rf.r.x, rf.r.y); 45162232bf4SGerd Hoffmann 45262232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, rf.resource_id); 45362232bf4SGerd Hoffmann if (!res) { 45462232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 45562232bf4SGerd Hoffmann __func__, rf.resource_id); 45662232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 45762232bf4SGerd Hoffmann return; 45862232bf4SGerd Hoffmann } 45962232bf4SGerd Hoffmann 46062232bf4SGerd Hoffmann if (rf.r.x > res->width || 46162232bf4SGerd Hoffmann rf.r.y > res->height || 46262232bf4SGerd Hoffmann rf.r.width > res->width || 46362232bf4SGerd Hoffmann rf.r.height > res->height || 46462232bf4SGerd Hoffmann rf.r.x + rf.r.width > res->width || 46562232bf4SGerd Hoffmann rf.r.y + rf.r.height > res->height) { 46662232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: flush bounds outside resource" 46762232bf4SGerd Hoffmann " bounds for resource %d: %d %d %d %d vs %d %d\n", 46862232bf4SGerd Hoffmann __func__, rf.resource_id, rf.r.x, rf.r.y, 46962232bf4SGerd Hoffmann rf.r.width, rf.r.height, res->width, res->height); 47062232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 47162232bf4SGerd Hoffmann return; 47262232bf4SGerd Hoffmann } 47362232bf4SGerd Hoffmann 47462232bf4SGerd Hoffmann pixman_region_init_rect(&flush_region, 47562232bf4SGerd Hoffmann rf.r.x, rf.r.y, rf.r.width, rf.r.height); 4762fe76055SMarc-André Lureau for (i = 0; i < g->conf.max_outputs; i++) { 47762232bf4SGerd Hoffmann struct virtio_gpu_scanout *scanout; 47862232bf4SGerd Hoffmann pixman_region16_t region, finalregion; 47962232bf4SGerd Hoffmann pixman_box16_t *extents; 48062232bf4SGerd Hoffmann 48162232bf4SGerd Hoffmann if (!(res->scanout_bitmask & (1 << i))) { 48262232bf4SGerd Hoffmann continue; 48362232bf4SGerd Hoffmann } 48462232bf4SGerd Hoffmann scanout = &g->scanout[i]; 48562232bf4SGerd Hoffmann 48662232bf4SGerd Hoffmann pixman_region_init(&finalregion); 48762232bf4SGerd Hoffmann pixman_region_init_rect(®ion, scanout->x, scanout->y, 48862232bf4SGerd Hoffmann scanout->width, scanout->height); 48962232bf4SGerd Hoffmann 49062232bf4SGerd Hoffmann pixman_region_intersect(&finalregion, &flush_region, ®ion); 49162232bf4SGerd Hoffmann pixman_region_translate(&finalregion, -scanout->x, -scanout->y); 49262232bf4SGerd Hoffmann extents = pixman_region_extents(&finalregion); 49362232bf4SGerd Hoffmann /* work out the area we need to update for each console */ 49462232bf4SGerd Hoffmann dpy_gfx_update(g->scanout[i].con, 49562232bf4SGerd Hoffmann extents->x1, extents->y1, 49662232bf4SGerd Hoffmann extents->x2 - extents->x1, 49762232bf4SGerd Hoffmann extents->y2 - extents->y1); 49862232bf4SGerd Hoffmann 49962232bf4SGerd Hoffmann pixman_region_fini(®ion); 50062232bf4SGerd Hoffmann pixman_region_fini(&finalregion); 50162232bf4SGerd Hoffmann } 50262232bf4SGerd Hoffmann pixman_region_fini(&flush_region); 50362232bf4SGerd Hoffmann } 50462232bf4SGerd Hoffmann 505fa06e5cbSGerd Hoffmann static void virtio_unref_resource(pixman_image_t *image, void *data) 506fa06e5cbSGerd Hoffmann { 507fa06e5cbSGerd Hoffmann pixman_image_unref(data); 508fa06e5cbSGerd Hoffmann } 509fa06e5cbSGerd Hoffmann 51062232bf4SGerd Hoffmann static void virtio_gpu_set_scanout(VirtIOGPU *g, 51162232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 51262232bf4SGerd Hoffmann { 51362232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 51462232bf4SGerd Hoffmann struct virtio_gpu_scanout *scanout; 51562232bf4SGerd Hoffmann pixman_format_code_t format; 51662232bf4SGerd Hoffmann uint32_t offset; 51762232bf4SGerd Hoffmann int bpp; 51862232bf4SGerd Hoffmann struct virtio_gpu_set_scanout ss; 51962232bf4SGerd Hoffmann 52062232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(ss); 52162232bf4SGerd Hoffmann trace_virtio_gpu_cmd_set_scanout(ss.scanout_id, ss.resource_id, 52262232bf4SGerd Hoffmann ss.r.width, ss.r.height, ss.r.x, ss.r.y); 52362232bf4SGerd Hoffmann 5242fe76055SMarc-André Lureau if (ss.scanout_id >= g->conf.max_outputs) { 525fe89fdebSMarc-André Lureau qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout id specified %d", 526fe89fdebSMarc-André Lureau __func__, ss.scanout_id); 527fe89fdebSMarc-André Lureau cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_SCANOUT_ID; 528fe89fdebSMarc-André Lureau return; 529fe89fdebSMarc-André Lureau } 530fe89fdebSMarc-André Lureau 53162232bf4SGerd Hoffmann g->enable = 1; 53262232bf4SGerd Hoffmann if (ss.resource_id == 0) { 53362232bf4SGerd Hoffmann scanout = &g->scanout[ss.scanout_id]; 53462232bf4SGerd Hoffmann if (scanout->resource_id) { 53562232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, scanout->resource_id); 53662232bf4SGerd Hoffmann if (res) { 53762232bf4SGerd Hoffmann res->scanout_bitmask &= ~(1 << ss.scanout_id); 53862232bf4SGerd Hoffmann } 53962232bf4SGerd Hoffmann } 540fe89fdebSMarc-André Lureau if (ss.scanout_id == 0) { 54162232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 54262232bf4SGerd Hoffmann "%s: illegal scanout id specified %d", 54362232bf4SGerd Hoffmann __func__, ss.scanout_id); 54462232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_SCANOUT_ID; 54562232bf4SGerd Hoffmann return; 54662232bf4SGerd Hoffmann } 54762232bf4SGerd Hoffmann dpy_gfx_replace_surface(g->scanout[ss.scanout_id].con, NULL); 54862232bf4SGerd Hoffmann scanout->ds = NULL; 54962232bf4SGerd Hoffmann scanout->width = 0; 55062232bf4SGerd Hoffmann scanout->height = 0; 55162232bf4SGerd Hoffmann return; 55262232bf4SGerd Hoffmann } 55362232bf4SGerd Hoffmann 55462232bf4SGerd Hoffmann /* create a surface for this scanout */ 55562232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, ss.resource_id); 55662232bf4SGerd Hoffmann if (!res) { 55762232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 55862232bf4SGerd Hoffmann __func__, ss.resource_id); 55962232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 56062232bf4SGerd Hoffmann return; 56162232bf4SGerd Hoffmann } 56262232bf4SGerd Hoffmann 56362232bf4SGerd Hoffmann if (ss.r.x > res->width || 56462232bf4SGerd Hoffmann ss.r.y > res->height || 56562232bf4SGerd Hoffmann ss.r.width > res->width || 56662232bf4SGerd Hoffmann ss.r.height > res->height || 56762232bf4SGerd Hoffmann ss.r.x + ss.r.width > res->width || 56862232bf4SGerd Hoffmann ss.r.y + ss.r.height > res->height) { 56962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout %d bounds for" 57062232bf4SGerd Hoffmann " resource %d, (%d,%d)+%d,%d vs %d %d\n", 57162232bf4SGerd Hoffmann __func__, ss.scanout_id, ss.resource_id, ss.r.x, ss.r.y, 57262232bf4SGerd Hoffmann ss.r.width, ss.r.height, res->width, res->height); 57362232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 57462232bf4SGerd Hoffmann return; 57562232bf4SGerd Hoffmann } 57662232bf4SGerd Hoffmann 57762232bf4SGerd Hoffmann scanout = &g->scanout[ss.scanout_id]; 57862232bf4SGerd Hoffmann 57962232bf4SGerd Hoffmann format = pixman_image_get_format(res->image); 58062232bf4SGerd Hoffmann bpp = (PIXMAN_FORMAT_BPP(format) + 7) / 8; 58162232bf4SGerd Hoffmann offset = (ss.r.x * bpp) + ss.r.y * pixman_image_get_stride(res->image); 58262232bf4SGerd Hoffmann if (!scanout->ds || surface_data(scanout->ds) 58362232bf4SGerd Hoffmann != ((uint8_t *)pixman_image_get_data(res->image) + offset) || 58462232bf4SGerd Hoffmann scanout->width != ss.r.width || 58562232bf4SGerd Hoffmann scanout->height != ss.r.height) { 586fa06e5cbSGerd Hoffmann pixman_image_t *rect; 587fa06e5cbSGerd Hoffmann void *ptr = (uint8_t *)pixman_image_get_data(res->image) + offset; 588fa06e5cbSGerd Hoffmann rect = pixman_image_create_bits(format, ss.r.width, ss.r.height, ptr, 589fa06e5cbSGerd Hoffmann pixman_image_get_stride(res->image)); 590fa06e5cbSGerd Hoffmann pixman_image_ref(res->image); 591fa06e5cbSGerd Hoffmann pixman_image_set_destroy_function(rect, virtio_unref_resource, 592fa06e5cbSGerd Hoffmann res->image); 59362232bf4SGerd Hoffmann /* realloc the surface ptr */ 594fa06e5cbSGerd Hoffmann scanout->ds = qemu_create_displaysurface_pixman(rect); 59562232bf4SGerd Hoffmann if (!scanout->ds) { 59662232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 59762232bf4SGerd Hoffmann return; 59862232bf4SGerd Hoffmann } 59962232bf4SGerd Hoffmann dpy_gfx_replace_surface(g->scanout[ss.scanout_id].con, scanout->ds); 60062232bf4SGerd Hoffmann } 60162232bf4SGerd Hoffmann 60262232bf4SGerd Hoffmann res->scanout_bitmask |= (1 << ss.scanout_id); 60362232bf4SGerd Hoffmann scanout->resource_id = ss.resource_id; 60462232bf4SGerd Hoffmann scanout->x = ss.r.x; 60562232bf4SGerd Hoffmann scanout->y = ss.r.y; 60662232bf4SGerd Hoffmann scanout->width = ss.r.width; 60762232bf4SGerd Hoffmann scanout->height = ss.r.height; 60862232bf4SGerd Hoffmann } 60962232bf4SGerd Hoffmann 61062232bf4SGerd Hoffmann int virtio_gpu_create_mapping_iov(struct virtio_gpu_resource_attach_backing *ab, 61162232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd, 612*0c244e50SGerd Hoffmann uint64_t **addr, struct iovec **iov) 61362232bf4SGerd Hoffmann { 61462232bf4SGerd Hoffmann struct virtio_gpu_mem_entry *ents; 61562232bf4SGerd Hoffmann size_t esize, s; 61662232bf4SGerd Hoffmann int i; 61762232bf4SGerd Hoffmann 61862232bf4SGerd Hoffmann if (ab->nr_entries > 16384) { 61962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 6202c84167bSGerd Hoffmann "%s: nr_entries is too big (%d > 16384)\n", 62162232bf4SGerd Hoffmann __func__, ab->nr_entries); 62262232bf4SGerd Hoffmann return -1; 62362232bf4SGerd Hoffmann } 62462232bf4SGerd Hoffmann 62562232bf4SGerd Hoffmann esize = sizeof(*ents) * ab->nr_entries; 62662232bf4SGerd Hoffmann ents = g_malloc(esize); 62762232bf4SGerd Hoffmann s = iov_to_buf(cmd->elem.out_sg, cmd->elem.out_num, 62862232bf4SGerd Hoffmann sizeof(*ab), ents, esize); 62962232bf4SGerd Hoffmann if (s != esize) { 63062232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 63162232bf4SGerd Hoffmann "%s: command data size incorrect %zu vs %zu\n", 63262232bf4SGerd Hoffmann __func__, s, esize); 63362232bf4SGerd Hoffmann g_free(ents); 63462232bf4SGerd Hoffmann return -1; 63562232bf4SGerd Hoffmann } 63662232bf4SGerd Hoffmann 63762232bf4SGerd Hoffmann *iov = g_malloc0(sizeof(struct iovec) * ab->nr_entries); 638*0c244e50SGerd Hoffmann if (addr) { 639*0c244e50SGerd Hoffmann *addr = g_malloc0(sizeof(uint64_t) * ab->nr_entries); 640*0c244e50SGerd Hoffmann } 64162232bf4SGerd Hoffmann for (i = 0; i < ab->nr_entries; i++) { 64262232bf4SGerd Hoffmann hwaddr len = ents[i].length; 64362232bf4SGerd Hoffmann (*iov)[i].iov_len = ents[i].length; 64462232bf4SGerd Hoffmann (*iov)[i].iov_base = cpu_physical_memory_map(ents[i].addr, &len, 1); 645*0c244e50SGerd Hoffmann if (addr) { 646*0c244e50SGerd Hoffmann (*addr)[i] = ents[i].addr; 647*0c244e50SGerd Hoffmann } 64862232bf4SGerd Hoffmann if (!(*iov)[i].iov_base || len != ents[i].length) { 64962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: failed to map MMIO memory for" 65062232bf4SGerd Hoffmann " resource %d element %d\n", 65162232bf4SGerd Hoffmann __func__, ab->resource_id, i); 65262232bf4SGerd Hoffmann virtio_gpu_cleanup_mapping_iov(*iov, i); 65362232bf4SGerd Hoffmann g_free(ents); 65462232bf4SGerd Hoffmann *iov = NULL; 655*0c244e50SGerd Hoffmann if (addr) { 656*0c244e50SGerd Hoffmann g_free(*addr); 657*0c244e50SGerd Hoffmann *addr = NULL; 658*0c244e50SGerd Hoffmann } 65962232bf4SGerd Hoffmann return -1; 66062232bf4SGerd Hoffmann } 66162232bf4SGerd Hoffmann } 66262232bf4SGerd Hoffmann g_free(ents); 66362232bf4SGerd Hoffmann return 0; 66462232bf4SGerd Hoffmann } 66562232bf4SGerd Hoffmann 66662232bf4SGerd Hoffmann void virtio_gpu_cleanup_mapping_iov(struct iovec *iov, uint32_t count) 66762232bf4SGerd Hoffmann { 66862232bf4SGerd Hoffmann int i; 66962232bf4SGerd Hoffmann 67062232bf4SGerd Hoffmann for (i = 0; i < count; i++) { 67162232bf4SGerd Hoffmann cpu_physical_memory_unmap(iov[i].iov_base, iov[i].iov_len, 1, 67262232bf4SGerd Hoffmann iov[i].iov_len); 67362232bf4SGerd Hoffmann } 6747f3be0f2SGerd Hoffmann g_free(iov); 67562232bf4SGerd Hoffmann } 67662232bf4SGerd Hoffmann 67762232bf4SGerd Hoffmann static void virtio_gpu_cleanup_mapping(struct virtio_gpu_simple_resource *res) 67862232bf4SGerd Hoffmann { 67962232bf4SGerd Hoffmann virtio_gpu_cleanup_mapping_iov(res->iov, res->iov_cnt); 68062232bf4SGerd Hoffmann res->iov = NULL; 68162232bf4SGerd Hoffmann res->iov_cnt = 0; 682*0c244e50SGerd Hoffmann g_free(res->addrs); 683*0c244e50SGerd Hoffmann res->addrs = NULL; 68462232bf4SGerd Hoffmann } 68562232bf4SGerd Hoffmann 68662232bf4SGerd Hoffmann static void 68762232bf4SGerd Hoffmann virtio_gpu_resource_attach_backing(VirtIOGPU *g, 68862232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 68962232bf4SGerd Hoffmann { 69062232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 69162232bf4SGerd Hoffmann struct virtio_gpu_resource_attach_backing ab; 69262232bf4SGerd Hoffmann int ret; 69362232bf4SGerd Hoffmann 69462232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(ab); 69562232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_back_attach(ab.resource_id); 69662232bf4SGerd Hoffmann 69762232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, ab.resource_id); 69862232bf4SGerd Hoffmann if (!res) { 69962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 70062232bf4SGerd Hoffmann __func__, ab.resource_id); 70162232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 70262232bf4SGerd Hoffmann return; 70362232bf4SGerd Hoffmann } 70462232bf4SGerd Hoffmann 705*0c244e50SGerd Hoffmann ret = virtio_gpu_create_mapping_iov(&ab, cmd, &res->addrs, &res->iov); 70662232bf4SGerd Hoffmann if (ret != 0) { 70762232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 70862232bf4SGerd Hoffmann return; 70962232bf4SGerd Hoffmann } 71062232bf4SGerd Hoffmann 71162232bf4SGerd Hoffmann res->iov_cnt = ab.nr_entries; 71262232bf4SGerd Hoffmann } 71362232bf4SGerd Hoffmann 71462232bf4SGerd Hoffmann static void 71562232bf4SGerd Hoffmann virtio_gpu_resource_detach_backing(VirtIOGPU *g, 71662232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 71762232bf4SGerd Hoffmann { 71862232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 71962232bf4SGerd Hoffmann struct virtio_gpu_resource_detach_backing detach; 72062232bf4SGerd Hoffmann 72162232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(detach); 72262232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_back_detach(detach.resource_id); 72362232bf4SGerd Hoffmann 72462232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, detach.resource_id); 72562232bf4SGerd Hoffmann if (!res || !res->iov) { 72662232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 72762232bf4SGerd Hoffmann __func__, detach.resource_id); 72862232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 72962232bf4SGerd Hoffmann return; 73062232bf4SGerd Hoffmann } 73162232bf4SGerd Hoffmann virtio_gpu_cleanup_mapping(res); 73262232bf4SGerd Hoffmann } 73362232bf4SGerd Hoffmann 73462232bf4SGerd Hoffmann static void virtio_gpu_simple_process_cmd(VirtIOGPU *g, 73562232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 73662232bf4SGerd Hoffmann { 73762232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(cmd->cmd_hdr); 73862232bf4SGerd Hoffmann 73962232bf4SGerd Hoffmann switch (cmd->cmd_hdr.type) { 74062232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_GET_DISPLAY_INFO: 74162232bf4SGerd Hoffmann virtio_gpu_get_display_info(g, cmd); 74262232bf4SGerd Hoffmann break; 74362232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_CREATE_2D: 74462232bf4SGerd Hoffmann virtio_gpu_resource_create_2d(g, cmd); 74562232bf4SGerd Hoffmann break; 74662232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_UNREF: 74762232bf4SGerd Hoffmann virtio_gpu_resource_unref(g, cmd); 74862232bf4SGerd Hoffmann break; 74962232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_FLUSH: 75062232bf4SGerd Hoffmann virtio_gpu_resource_flush(g, cmd); 75162232bf4SGerd Hoffmann break; 75262232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_TRANSFER_TO_HOST_2D: 75362232bf4SGerd Hoffmann virtio_gpu_transfer_to_host_2d(g, cmd); 75462232bf4SGerd Hoffmann break; 75562232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_SET_SCANOUT: 75662232bf4SGerd Hoffmann virtio_gpu_set_scanout(g, cmd); 75762232bf4SGerd Hoffmann break; 75862232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_ATTACH_BACKING: 75962232bf4SGerd Hoffmann virtio_gpu_resource_attach_backing(g, cmd); 76062232bf4SGerd Hoffmann break; 76162232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_DETACH_BACKING: 76262232bf4SGerd Hoffmann virtio_gpu_resource_detach_backing(g, cmd); 76362232bf4SGerd Hoffmann break; 76462232bf4SGerd Hoffmann default: 76562232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 76662232bf4SGerd Hoffmann break; 76762232bf4SGerd Hoffmann } 76862232bf4SGerd Hoffmann if (!cmd->finished) { 76962232bf4SGerd Hoffmann virtio_gpu_ctrl_response_nodata(g, cmd, cmd->error ? cmd->error : 77062232bf4SGerd Hoffmann VIRTIO_GPU_RESP_OK_NODATA); 77162232bf4SGerd Hoffmann } 77262232bf4SGerd Hoffmann } 77362232bf4SGerd Hoffmann 77462232bf4SGerd Hoffmann static void virtio_gpu_handle_ctrl_cb(VirtIODevice *vdev, VirtQueue *vq) 77562232bf4SGerd Hoffmann { 77662232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 77762232bf4SGerd Hoffmann qemu_bh_schedule(g->ctrl_bh); 77862232bf4SGerd Hoffmann } 77962232bf4SGerd Hoffmann 78062232bf4SGerd Hoffmann static void virtio_gpu_handle_cursor_cb(VirtIODevice *vdev, VirtQueue *vq) 78162232bf4SGerd Hoffmann { 78262232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 78362232bf4SGerd Hoffmann qemu_bh_schedule(g->cursor_bh); 78462232bf4SGerd Hoffmann } 78562232bf4SGerd Hoffmann 7860c55a1cfSGerd Hoffmann void virtio_gpu_process_cmdq(VirtIOGPU *g) 7873eb769fdSGerd Hoffmann { 7883eb769fdSGerd Hoffmann struct virtio_gpu_ctrl_command *cmd; 7893eb769fdSGerd Hoffmann 7903eb769fdSGerd Hoffmann while (!QTAILQ_EMPTY(&g->cmdq)) { 7913eb769fdSGerd Hoffmann cmd = QTAILQ_FIRST(&g->cmdq); 7923eb769fdSGerd Hoffmann 7933eb769fdSGerd Hoffmann /* process command */ 7943eb769fdSGerd Hoffmann VIRGL(g, virtio_gpu_virgl_process_cmd, virtio_gpu_simple_process_cmd, 7953eb769fdSGerd Hoffmann g, cmd); 7960c55a1cfSGerd Hoffmann if (cmd->waiting) { 7970c55a1cfSGerd Hoffmann break; 7980c55a1cfSGerd Hoffmann } 7993eb769fdSGerd Hoffmann QTAILQ_REMOVE(&g->cmdq, cmd, next); 8003eb769fdSGerd Hoffmann if (virtio_gpu_stats_enabled(g->conf)) { 8013eb769fdSGerd Hoffmann g->stats.requests++; 8023eb769fdSGerd Hoffmann } 8033eb769fdSGerd Hoffmann 8043eb769fdSGerd Hoffmann if (!cmd->finished) { 8053eb769fdSGerd Hoffmann QTAILQ_INSERT_TAIL(&g->fenceq, cmd, next); 8063eb769fdSGerd Hoffmann g->inflight++; 8073eb769fdSGerd Hoffmann if (virtio_gpu_stats_enabled(g->conf)) { 8083eb769fdSGerd Hoffmann if (g->stats.max_inflight < g->inflight) { 8093eb769fdSGerd Hoffmann g->stats.max_inflight = g->inflight; 8103eb769fdSGerd Hoffmann } 8113eb769fdSGerd Hoffmann fprintf(stderr, "inflight: %3d (+)\r", g->inflight); 8123eb769fdSGerd Hoffmann } 8133eb769fdSGerd Hoffmann } else { 8143eb769fdSGerd Hoffmann g_free(cmd); 8153eb769fdSGerd Hoffmann } 8163eb769fdSGerd Hoffmann } 8173eb769fdSGerd Hoffmann } 8183eb769fdSGerd Hoffmann 81962232bf4SGerd Hoffmann static void virtio_gpu_handle_ctrl(VirtIODevice *vdev, VirtQueue *vq) 82062232bf4SGerd Hoffmann { 82162232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 82262232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd; 82362232bf4SGerd Hoffmann 82462232bf4SGerd Hoffmann if (!virtio_queue_ready(vq)) { 82562232bf4SGerd Hoffmann return; 82662232bf4SGerd Hoffmann } 82762232bf4SGerd Hoffmann 8289d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 8299d9e1521SGerd Hoffmann if (!g->renderer_inited && g->use_virgl_renderer) { 8309d9e1521SGerd Hoffmann virtio_gpu_virgl_init(g); 8319d9e1521SGerd Hoffmann g->renderer_inited = true; 8329d9e1521SGerd Hoffmann } 8339d9e1521SGerd Hoffmann #endif 8349d9e1521SGerd Hoffmann 83551b19ebeSPaolo Bonzini cmd = virtqueue_pop(vq, sizeof(struct virtio_gpu_ctrl_command)); 83651b19ebeSPaolo Bonzini while (cmd) { 83762232bf4SGerd Hoffmann cmd->vq = vq; 83862232bf4SGerd Hoffmann cmd->error = 0; 83962232bf4SGerd Hoffmann cmd->finished = false; 8403eb769fdSGerd Hoffmann cmd->waiting = false; 8413eb769fdSGerd Hoffmann QTAILQ_INSERT_TAIL(&g->cmdq, cmd, next); 84251b19ebeSPaolo Bonzini cmd = virtqueue_pop(vq, sizeof(struct virtio_gpu_ctrl_command)); 84362232bf4SGerd Hoffmann } 8449d9e1521SGerd Hoffmann 8453eb769fdSGerd Hoffmann virtio_gpu_process_cmdq(g); 8463eb769fdSGerd Hoffmann 8479d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 8489d9e1521SGerd Hoffmann if (g->use_virgl_renderer) { 8499d9e1521SGerd Hoffmann virtio_gpu_virgl_fence_poll(g); 8509d9e1521SGerd Hoffmann } 8519d9e1521SGerd Hoffmann #endif 85262232bf4SGerd Hoffmann } 85362232bf4SGerd Hoffmann 85462232bf4SGerd Hoffmann static void virtio_gpu_ctrl_bh(void *opaque) 85562232bf4SGerd Hoffmann { 85662232bf4SGerd Hoffmann VirtIOGPU *g = opaque; 85762232bf4SGerd Hoffmann virtio_gpu_handle_ctrl(&g->parent_obj, g->ctrl_vq); 85862232bf4SGerd Hoffmann } 85962232bf4SGerd Hoffmann 86062232bf4SGerd Hoffmann static void virtio_gpu_handle_cursor(VirtIODevice *vdev, VirtQueue *vq) 86162232bf4SGerd Hoffmann { 86262232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 86351b19ebeSPaolo Bonzini VirtQueueElement *elem; 86462232bf4SGerd Hoffmann size_t s; 86562232bf4SGerd Hoffmann struct virtio_gpu_update_cursor cursor_info; 86662232bf4SGerd Hoffmann 86762232bf4SGerd Hoffmann if (!virtio_queue_ready(vq)) { 86862232bf4SGerd Hoffmann return; 86962232bf4SGerd Hoffmann } 87051b19ebeSPaolo Bonzini for (;;) { 87151b19ebeSPaolo Bonzini elem = virtqueue_pop(vq, sizeof(VirtQueueElement)); 87251b19ebeSPaolo Bonzini if (!elem) { 87351b19ebeSPaolo Bonzini break; 87451b19ebeSPaolo Bonzini } 87551b19ebeSPaolo Bonzini 87651b19ebeSPaolo Bonzini s = iov_to_buf(elem->out_sg, elem->out_num, 0, 87762232bf4SGerd Hoffmann &cursor_info, sizeof(cursor_info)); 87862232bf4SGerd Hoffmann if (s != sizeof(cursor_info)) { 87962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 88062232bf4SGerd Hoffmann "%s: cursor size incorrect %zu vs %zu\n", 88162232bf4SGerd Hoffmann __func__, s, sizeof(cursor_info)); 88262232bf4SGerd Hoffmann } else { 88362232bf4SGerd Hoffmann update_cursor(g, &cursor_info); 88462232bf4SGerd Hoffmann } 88551b19ebeSPaolo Bonzini virtqueue_push(vq, elem, 0); 88662232bf4SGerd Hoffmann virtio_notify(vdev, vq); 88751b19ebeSPaolo Bonzini g_free(elem); 88862232bf4SGerd Hoffmann } 88962232bf4SGerd Hoffmann } 89062232bf4SGerd Hoffmann 89162232bf4SGerd Hoffmann static void virtio_gpu_cursor_bh(void *opaque) 89262232bf4SGerd Hoffmann { 89362232bf4SGerd Hoffmann VirtIOGPU *g = opaque; 89462232bf4SGerd Hoffmann virtio_gpu_handle_cursor(&g->parent_obj, g->cursor_vq); 89562232bf4SGerd Hoffmann } 89662232bf4SGerd Hoffmann 89762232bf4SGerd Hoffmann static void virtio_gpu_invalidate_display(void *opaque) 89862232bf4SGerd Hoffmann { 89962232bf4SGerd Hoffmann } 90062232bf4SGerd Hoffmann 90162232bf4SGerd Hoffmann static void virtio_gpu_update_display(void *opaque) 90262232bf4SGerd Hoffmann { 90362232bf4SGerd Hoffmann } 90462232bf4SGerd Hoffmann 90562232bf4SGerd Hoffmann static void virtio_gpu_text_update(void *opaque, console_ch_t *chardata) 90662232bf4SGerd Hoffmann { 90762232bf4SGerd Hoffmann } 90862232bf4SGerd Hoffmann 90962232bf4SGerd Hoffmann static int virtio_gpu_ui_info(void *opaque, uint32_t idx, QemuUIInfo *info) 91062232bf4SGerd Hoffmann { 91162232bf4SGerd Hoffmann VirtIOGPU *g = opaque; 91262232bf4SGerd Hoffmann 9136b860806SMarc-André Lureau if (idx >= g->conf.max_outputs) { 91462232bf4SGerd Hoffmann return -1; 91562232bf4SGerd Hoffmann } 91662232bf4SGerd Hoffmann 91762232bf4SGerd Hoffmann g->req_state[idx].x = info->xoff; 91862232bf4SGerd Hoffmann g->req_state[idx].y = info->yoff; 91962232bf4SGerd Hoffmann g->req_state[idx].width = info->width; 92062232bf4SGerd Hoffmann g->req_state[idx].height = info->height; 92162232bf4SGerd Hoffmann 92262232bf4SGerd Hoffmann if (info->width && info->height) { 92362232bf4SGerd Hoffmann g->enabled_output_bitmask |= (1 << idx); 92462232bf4SGerd Hoffmann } else { 92562232bf4SGerd Hoffmann g->enabled_output_bitmask &= ~(1 << idx); 92662232bf4SGerd Hoffmann } 92762232bf4SGerd Hoffmann 92862232bf4SGerd Hoffmann /* send event to guest */ 92962232bf4SGerd Hoffmann virtio_gpu_notify_event(g, VIRTIO_GPU_EVENT_DISPLAY); 93062232bf4SGerd Hoffmann return 0; 93162232bf4SGerd Hoffmann } 93262232bf4SGerd Hoffmann 933321c9adbSGerd Hoffmann static void virtio_gpu_gl_block(void *opaque, bool block) 934321c9adbSGerd Hoffmann { 935321c9adbSGerd Hoffmann VirtIOGPU *g = opaque; 936321c9adbSGerd Hoffmann 937321c9adbSGerd Hoffmann g->renderer_blocked = block; 938321c9adbSGerd Hoffmann if (!block) { 939321c9adbSGerd Hoffmann virtio_gpu_process_cmdq(g); 940321c9adbSGerd Hoffmann } 941321c9adbSGerd Hoffmann } 942321c9adbSGerd Hoffmann 94362232bf4SGerd Hoffmann const GraphicHwOps virtio_gpu_ops = { 94462232bf4SGerd Hoffmann .invalidate = virtio_gpu_invalidate_display, 94562232bf4SGerd Hoffmann .gfx_update = virtio_gpu_update_display, 94662232bf4SGerd Hoffmann .text_update = virtio_gpu_text_update, 94762232bf4SGerd Hoffmann .ui_info = virtio_gpu_ui_info, 948321c9adbSGerd Hoffmann .gl_block = virtio_gpu_gl_block, 94962232bf4SGerd Hoffmann }; 95062232bf4SGerd Hoffmann 951*0c244e50SGerd Hoffmann static const VMStateDescription vmstate_virtio_gpu_scanout = { 952*0c244e50SGerd Hoffmann .name = "virtio-gpu-one-scanout", 953*0c244e50SGerd Hoffmann .version_id = 1, 954*0c244e50SGerd Hoffmann .fields = (VMStateField[]) { 955*0c244e50SGerd Hoffmann VMSTATE_UINT32(resource_id, struct virtio_gpu_scanout), 956*0c244e50SGerd Hoffmann VMSTATE_UINT32(width, struct virtio_gpu_scanout), 957*0c244e50SGerd Hoffmann VMSTATE_UINT32(height, struct virtio_gpu_scanout), 958*0c244e50SGerd Hoffmann VMSTATE_INT32(x, struct virtio_gpu_scanout), 959*0c244e50SGerd Hoffmann VMSTATE_INT32(y, struct virtio_gpu_scanout), 960*0c244e50SGerd Hoffmann VMSTATE_UINT32(cursor.resource_id, struct virtio_gpu_scanout), 961*0c244e50SGerd Hoffmann VMSTATE_UINT32(cursor.hot_x, struct virtio_gpu_scanout), 962*0c244e50SGerd Hoffmann VMSTATE_UINT32(cursor.hot_y, struct virtio_gpu_scanout), 963*0c244e50SGerd Hoffmann VMSTATE_UINT32(cursor.pos.x, struct virtio_gpu_scanout), 964*0c244e50SGerd Hoffmann VMSTATE_UINT32(cursor.pos.y, struct virtio_gpu_scanout), 965*0c244e50SGerd Hoffmann VMSTATE_END_OF_LIST() 966*0c244e50SGerd Hoffmann }, 967*0c244e50SGerd Hoffmann }; 968*0c244e50SGerd Hoffmann 969*0c244e50SGerd Hoffmann static const VMStateDescription vmstate_virtio_gpu_scanouts = { 970*0c244e50SGerd Hoffmann .name = "virtio-gpu-scanouts", 971*0c244e50SGerd Hoffmann .version_id = 1, 972*0c244e50SGerd Hoffmann .fields = (VMStateField[]) { 973*0c244e50SGerd Hoffmann VMSTATE_INT32(enable, struct VirtIOGPU), 974*0c244e50SGerd Hoffmann VMSTATE_UINT32_EQUAL(conf.max_outputs, struct VirtIOGPU), 975*0c244e50SGerd Hoffmann VMSTATE_STRUCT_VARRAY_UINT32(scanout, struct VirtIOGPU, 976*0c244e50SGerd Hoffmann conf.max_outputs, 1, 977*0c244e50SGerd Hoffmann vmstate_virtio_gpu_scanout, 978*0c244e50SGerd Hoffmann struct virtio_gpu_scanout), 979*0c244e50SGerd Hoffmann VMSTATE_END_OF_LIST() 980*0c244e50SGerd Hoffmann }, 981*0c244e50SGerd Hoffmann }; 982*0c244e50SGerd Hoffmann 983fa49e465SGerd Hoffmann static const VMStateDescription vmstate_virtio_gpu_unmigratable = { 984*0c244e50SGerd Hoffmann .name = "virtio-gpu-with-virgl", 985fa49e465SGerd Hoffmann .unmigratable = 1, 986fa49e465SGerd Hoffmann }; 987fa49e465SGerd Hoffmann 988*0c244e50SGerd Hoffmann static void virtio_gpu_save(QEMUFile *f, void *opaque) 989*0c244e50SGerd Hoffmann { 990*0c244e50SGerd Hoffmann VirtIOGPU *g = opaque; 991*0c244e50SGerd Hoffmann VirtIODevice *vdev = VIRTIO_DEVICE(g); 992*0c244e50SGerd Hoffmann struct virtio_gpu_simple_resource *res; 993*0c244e50SGerd Hoffmann int i; 994*0c244e50SGerd Hoffmann 995*0c244e50SGerd Hoffmann virtio_save(vdev, f); 996*0c244e50SGerd Hoffmann 997*0c244e50SGerd Hoffmann /* in 2d mode we should never find unprocessed commands here */ 998*0c244e50SGerd Hoffmann assert(QTAILQ_EMPTY(&g->cmdq)); 999*0c244e50SGerd Hoffmann 1000*0c244e50SGerd Hoffmann QTAILQ_FOREACH(res, &g->reslist, next) { 1001*0c244e50SGerd Hoffmann qemu_put_be32(f, res->resource_id); 1002*0c244e50SGerd Hoffmann qemu_put_be32(f, res->width); 1003*0c244e50SGerd Hoffmann qemu_put_be32(f, res->height); 1004*0c244e50SGerd Hoffmann qemu_put_be32(f, res->format); 1005*0c244e50SGerd Hoffmann qemu_put_be32(f, res->iov_cnt); 1006*0c244e50SGerd Hoffmann for (i = 0; i < res->iov_cnt; i++) { 1007*0c244e50SGerd Hoffmann qemu_put_be64(f, res->addrs[i]); 1008*0c244e50SGerd Hoffmann qemu_put_be32(f, res->iov[i].iov_len); 1009*0c244e50SGerd Hoffmann } 1010*0c244e50SGerd Hoffmann qemu_put_buffer(f, (void *)pixman_image_get_data(res->image), 1011*0c244e50SGerd Hoffmann pixman_image_get_stride(res->image) * res->height); 1012*0c244e50SGerd Hoffmann } 1013*0c244e50SGerd Hoffmann qemu_put_be32(f, 0); /* end of list */ 1014*0c244e50SGerd Hoffmann 1015*0c244e50SGerd Hoffmann vmstate_save_state(f, &vmstate_virtio_gpu_scanouts, g, NULL); 1016*0c244e50SGerd Hoffmann } 1017*0c244e50SGerd Hoffmann 1018*0c244e50SGerd Hoffmann static int virtio_gpu_load(QEMUFile *f, void *opaque, int version_id) 1019*0c244e50SGerd Hoffmann { 1020*0c244e50SGerd Hoffmann VirtIOGPU *g = opaque; 1021*0c244e50SGerd Hoffmann VirtIODevice *vdev = VIRTIO_DEVICE(g); 1022*0c244e50SGerd Hoffmann struct virtio_gpu_simple_resource *res; 1023*0c244e50SGerd Hoffmann struct virtio_gpu_scanout *scanout; 1024*0c244e50SGerd Hoffmann uint32_t resource_id, pformat; 1025*0c244e50SGerd Hoffmann int i, ret; 1026*0c244e50SGerd Hoffmann 1027*0c244e50SGerd Hoffmann if (version_id != VIRTIO_GPU_VM_VERSION) { 1028*0c244e50SGerd Hoffmann return -EINVAL; 1029*0c244e50SGerd Hoffmann } 1030*0c244e50SGerd Hoffmann 1031*0c244e50SGerd Hoffmann ret = virtio_load(vdev, f, version_id); 1032*0c244e50SGerd Hoffmann if (ret) { 1033*0c244e50SGerd Hoffmann return ret; 1034*0c244e50SGerd Hoffmann } 1035*0c244e50SGerd Hoffmann 1036*0c244e50SGerd Hoffmann resource_id = qemu_get_be32(f); 1037*0c244e50SGerd Hoffmann while (resource_id != 0) { 1038*0c244e50SGerd Hoffmann res = g_new0(struct virtio_gpu_simple_resource, 1); 1039*0c244e50SGerd Hoffmann res->resource_id = resource_id; 1040*0c244e50SGerd Hoffmann res->width = qemu_get_be32(f); 1041*0c244e50SGerd Hoffmann res->height = qemu_get_be32(f); 1042*0c244e50SGerd Hoffmann res->format = qemu_get_be32(f); 1043*0c244e50SGerd Hoffmann res->iov_cnt = qemu_get_be32(f); 1044*0c244e50SGerd Hoffmann 1045*0c244e50SGerd Hoffmann /* allocate */ 1046*0c244e50SGerd Hoffmann pformat = get_pixman_format(res->format); 1047*0c244e50SGerd Hoffmann if (!pformat) { 1048*0c244e50SGerd Hoffmann return -EINVAL; 1049*0c244e50SGerd Hoffmann } 1050*0c244e50SGerd Hoffmann res->image = pixman_image_create_bits(pformat, 1051*0c244e50SGerd Hoffmann res->width, res->height, 1052*0c244e50SGerd Hoffmann NULL, 0); 1053*0c244e50SGerd Hoffmann if (!res->image) { 1054*0c244e50SGerd Hoffmann return -EINVAL; 1055*0c244e50SGerd Hoffmann } 1056*0c244e50SGerd Hoffmann 1057*0c244e50SGerd Hoffmann res->addrs = g_new(uint64_t, res->iov_cnt); 1058*0c244e50SGerd Hoffmann res->iov = g_new(struct iovec, res->iov_cnt); 1059*0c244e50SGerd Hoffmann 1060*0c244e50SGerd Hoffmann /* read data */ 1061*0c244e50SGerd Hoffmann for (i = 0; i < res->iov_cnt; i++) { 1062*0c244e50SGerd Hoffmann res->addrs[i] = qemu_get_be64(f); 1063*0c244e50SGerd Hoffmann res->iov[i].iov_len = qemu_get_be32(f); 1064*0c244e50SGerd Hoffmann } 1065*0c244e50SGerd Hoffmann qemu_get_buffer(f, (void *)pixman_image_get_data(res->image), 1066*0c244e50SGerd Hoffmann pixman_image_get_stride(res->image) * res->height); 1067*0c244e50SGerd Hoffmann 1068*0c244e50SGerd Hoffmann /* restore mapping */ 1069*0c244e50SGerd Hoffmann for (i = 0; i < res->iov_cnt; i++) { 1070*0c244e50SGerd Hoffmann hwaddr len = res->iov[i].iov_len; 1071*0c244e50SGerd Hoffmann res->iov[i].iov_base = 1072*0c244e50SGerd Hoffmann cpu_physical_memory_map(res->addrs[i], &len, 1); 1073*0c244e50SGerd Hoffmann if (!res->iov[i].iov_base || len != res->iov[i].iov_len) { 1074*0c244e50SGerd Hoffmann return -EINVAL; 1075*0c244e50SGerd Hoffmann } 1076*0c244e50SGerd Hoffmann } 1077*0c244e50SGerd Hoffmann 1078*0c244e50SGerd Hoffmann QTAILQ_INSERT_HEAD(&g->reslist, res, next); 1079*0c244e50SGerd Hoffmann 1080*0c244e50SGerd Hoffmann resource_id = qemu_get_be32(f); 1081*0c244e50SGerd Hoffmann } 1082*0c244e50SGerd Hoffmann 1083*0c244e50SGerd Hoffmann /* load & apply scanout state */ 1084*0c244e50SGerd Hoffmann vmstate_load_state(f, &vmstate_virtio_gpu_scanouts, g, 1); 1085*0c244e50SGerd Hoffmann for (i = 0; i < g->conf.max_outputs; i++) { 1086*0c244e50SGerd Hoffmann scanout = &g->scanout[i]; 1087*0c244e50SGerd Hoffmann if (!scanout->resource_id) { 1088*0c244e50SGerd Hoffmann continue; 1089*0c244e50SGerd Hoffmann } 1090*0c244e50SGerd Hoffmann res = virtio_gpu_find_resource(g, scanout->resource_id); 1091*0c244e50SGerd Hoffmann if (!res) { 1092*0c244e50SGerd Hoffmann return -EINVAL; 1093*0c244e50SGerd Hoffmann } 1094*0c244e50SGerd Hoffmann scanout->ds = qemu_create_displaysurface_pixman(res->image); 1095*0c244e50SGerd Hoffmann if (!scanout->ds) { 1096*0c244e50SGerd Hoffmann return -EINVAL; 1097*0c244e50SGerd Hoffmann } 1098*0c244e50SGerd Hoffmann 1099*0c244e50SGerd Hoffmann dpy_gfx_replace_surface(scanout->con, scanout->ds); 1100*0c244e50SGerd Hoffmann dpy_gfx_update(scanout->con, 0, 0, scanout->width, scanout->height); 1101*0c244e50SGerd Hoffmann update_cursor(g, &scanout->cursor); 1102*0c244e50SGerd Hoffmann res->scanout_bitmask |= (1 << i); 1103*0c244e50SGerd Hoffmann } 1104*0c244e50SGerd Hoffmann 1105*0c244e50SGerd Hoffmann return 0; 1106*0c244e50SGerd Hoffmann } 1107*0c244e50SGerd Hoffmann 110862232bf4SGerd Hoffmann static void virtio_gpu_device_realize(DeviceState *qdev, Error **errp) 110962232bf4SGerd Hoffmann { 111062232bf4SGerd Hoffmann VirtIODevice *vdev = VIRTIO_DEVICE(qdev); 111162232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(qdev); 11129d9e1521SGerd Hoffmann bool have_virgl; 111362232bf4SGerd Hoffmann int i; 111462232bf4SGerd Hoffmann 1115acfc4846SMarc-André Lureau if (g->conf.max_outputs > VIRTIO_GPU_MAX_SCANOUTS) { 1116acfc4846SMarc-André Lureau error_setg(errp, "invalid max_outputs > %d", VIRTIO_GPU_MAX_SCANOUTS); 11175e3d741cSMarc-André Lureau return; 11185e3d741cSMarc-André Lureau } 11195e3d741cSMarc-André Lureau 112062232bf4SGerd Hoffmann g->config_size = sizeof(struct virtio_gpu_config); 112162232bf4SGerd Hoffmann g->virtio_config.num_scanouts = g->conf.max_outputs; 112262232bf4SGerd Hoffmann virtio_init(VIRTIO_DEVICE(g), "virtio-gpu", VIRTIO_ID_GPU, 112362232bf4SGerd Hoffmann g->config_size); 112462232bf4SGerd Hoffmann 112562232bf4SGerd Hoffmann g->req_state[0].width = 1024; 112662232bf4SGerd Hoffmann g->req_state[0].height = 768; 112762232bf4SGerd Hoffmann 11289d9e1521SGerd Hoffmann g->use_virgl_renderer = false; 11299d9e1521SGerd Hoffmann #if !defined(CONFIG_VIRGL) || defined(HOST_WORDS_BIGENDIAN) 11309d9e1521SGerd Hoffmann have_virgl = false; 11319d9e1521SGerd Hoffmann #else 11329d9e1521SGerd Hoffmann have_virgl = display_opengl; 11339d9e1521SGerd Hoffmann #endif 11349d9e1521SGerd Hoffmann if (!have_virgl) { 11359d9e1521SGerd Hoffmann g->conf.flags &= ~(1 << VIRTIO_GPU_FLAG_VIRGL_ENABLED); 11369d9e1521SGerd Hoffmann } 11379d9e1521SGerd Hoffmann 11389d9e1521SGerd Hoffmann if (virtio_gpu_virgl_enabled(g->conf)) { 11399d9e1521SGerd Hoffmann /* use larger control queue in 3d mode */ 11409d9e1521SGerd Hoffmann g->ctrl_vq = virtio_add_queue(vdev, 256, virtio_gpu_handle_ctrl_cb); 11419d9e1521SGerd Hoffmann g->cursor_vq = virtio_add_queue(vdev, 16, virtio_gpu_handle_cursor_cb); 11429d9e1521SGerd Hoffmann g->virtio_config.num_capsets = 1; 11439d9e1521SGerd Hoffmann } else { 114462232bf4SGerd Hoffmann g->ctrl_vq = virtio_add_queue(vdev, 64, virtio_gpu_handle_ctrl_cb); 114562232bf4SGerd Hoffmann g->cursor_vq = virtio_add_queue(vdev, 16, virtio_gpu_handle_cursor_cb); 11469d9e1521SGerd Hoffmann } 114762232bf4SGerd Hoffmann 114862232bf4SGerd Hoffmann g->ctrl_bh = qemu_bh_new(virtio_gpu_ctrl_bh, g); 114962232bf4SGerd Hoffmann g->cursor_bh = qemu_bh_new(virtio_gpu_cursor_bh, g); 115062232bf4SGerd Hoffmann QTAILQ_INIT(&g->reslist); 11513eb769fdSGerd Hoffmann QTAILQ_INIT(&g->cmdq); 115262232bf4SGerd Hoffmann QTAILQ_INIT(&g->fenceq); 115362232bf4SGerd Hoffmann 115462232bf4SGerd Hoffmann g->enabled_output_bitmask = 1; 115562232bf4SGerd Hoffmann g->qdev = qdev; 115662232bf4SGerd Hoffmann 115762232bf4SGerd Hoffmann for (i = 0; i < g->conf.max_outputs; i++) { 115862232bf4SGerd Hoffmann g->scanout[i].con = 115962232bf4SGerd Hoffmann graphic_console_init(DEVICE(g), i, &virtio_gpu_ops, g); 116062232bf4SGerd Hoffmann if (i > 0) { 116162232bf4SGerd Hoffmann dpy_gfx_replace_surface(g->scanout[i].con, NULL); 116262232bf4SGerd Hoffmann } 116362232bf4SGerd Hoffmann } 1164fa49e465SGerd Hoffmann 1165*0c244e50SGerd Hoffmann if (virtio_gpu_virgl_enabled(g->conf)) { 1166fa49e465SGerd Hoffmann vmstate_register(qdev, -1, &vmstate_virtio_gpu_unmigratable, g); 1167*0c244e50SGerd Hoffmann } else { 1168*0c244e50SGerd Hoffmann register_savevm(qdev, "virtio-gpu", -1, VIRTIO_GPU_VM_VERSION, 1169*0c244e50SGerd Hoffmann virtio_gpu_save, virtio_gpu_load, g); 1170*0c244e50SGerd Hoffmann } 117162232bf4SGerd Hoffmann } 117262232bf4SGerd Hoffmann 117362232bf4SGerd Hoffmann static void virtio_gpu_instance_init(Object *obj) 117462232bf4SGerd Hoffmann { 117562232bf4SGerd Hoffmann } 117662232bf4SGerd Hoffmann 117762232bf4SGerd Hoffmann static void virtio_gpu_reset(VirtIODevice *vdev) 117862232bf4SGerd Hoffmann { 117962232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 118062232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res, *tmp; 118162232bf4SGerd Hoffmann int i; 118262232bf4SGerd Hoffmann 118362232bf4SGerd Hoffmann g->enable = 0; 118462232bf4SGerd Hoffmann 118562232bf4SGerd Hoffmann QTAILQ_FOREACH_SAFE(res, &g->reslist, next, tmp) { 118662232bf4SGerd Hoffmann virtio_gpu_resource_destroy(g, res); 118762232bf4SGerd Hoffmann } 118862232bf4SGerd Hoffmann for (i = 0; i < g->conf.max_outputs; i++) { 118962232bf4SGerd Hoffmann #if 0 119062232bf4SGerd Hoffmann g->req_state[i].x = 0; 119162232bf4SGerd Hoffmann g->req_state[i].y = 0; 119262232bf4SGerd Hoffmann if (i == 0) { 119362232bf4SGerd Hoffmann g->req_state[0].width = 1024; 119462232bf4SGerd Hoffmann g->req_state[0].height = 768; 119562232bf4SGerd Hoffmann } else { 119662232bf4SGerd Hoffmann g->req_state[i].width = 0; 119762232bf4SGerd Hoffmann g->req_state[i].height = 0; 119862232bf4SGerd Hoffmann } 119962232bf4SGerd Hoffmann #endif 120062232bf4SGerd Hoffmann g->scanout[i].resource_id = 0; 120162232bf4SGerd Hoffmann g->scanout[i].width = 0; 120262232bf4SGerd Hoffmann g->scanout[i].height = 0; 120362232bf4SGerd Hoffmann g->scanout[i].x = 0; 120462232bf4SGerd Hoffmann g->scanout[i].y = 0; 120562232bf4SGerd Hoffmann g->scanout[i].ds = NULL; 120662232bf4SGerd Hoffmann } 120762232bf4SGerd Hoffmann g->enabled_output_bitmask = 1; 12089d9e1521SGerd Hoffmann 12099d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 12109d9e1521SGerd Hoffmann if (g->use_virgl_renderer) { 12119d9e1521SGerd Hoffmann virtio_gpu_virgl_reset(g); 12129d9e1521SGerd Hoffmann g->use_virgl_renderer = 0; 12139d9e1521SGerd Hoffmann } 12149d9e1521SGerd Hoffmann #endif 121562232bf4SGerd Hoffmann } 121662232bf4SGerd Hoffmann 121762232bf4SGerd Hoffmann static Property virtio_gpu_properties[] = { 1218b3409a31SGerd Hoffmann DEFINE_PROP_UINT32("max_outputs", VirtIOGPU, conf.max_outputs, 1), 12199d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 12209d9e1521SGerd Hoffmann DEFINE_PROP_BIT("virgl", VirtIOGPU, conf.flags, 12219d9e1521SGerd Hoffmann VIRTIO_GPU_FLAG_VIRGL_ENABLED, true), 12229d9e1521SGerd Hoffmann DEFINE_PROP_BIT("stats", VirtIOGPU, conf.flags, 12239d9e1521SGerd Hoffmann VIRTIO_GPU_FLAG_STATS_ENABLED, false), 12249d9e1521SGerd Hoffmann #endif 122562232bf4SGerd Hoffmann DEFINE_PROP_END_OF_LIST(), 122662232bf4SGerd Hoffmann }; 122762232bf4SGerd Hoffmann 122862232bf4SGerd Hoffmann static void virtio_gpu_class_init(ObjectClass *klass, void *data) 122962232bf4SGerd Hoffmann { 123062232bf4SGerd Hoffmann DeviceClass *dc = DEVICE_CLASS(klass); 123162232bf4SGerd Hoffmann VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass); 123262232bf4SGerd Hoffmann 123362232bf4SGerd Hoffmann vdc->realize = virtio_gpu_device_realize; 123462232bf4SGerd Hoffmann vdc->get_config = virtio_gpu_get_config; 123562232bf4SGerd Hoffmann vdc->set_config = virtio_gpu_set_config; 123662232bf4SGerd Hoffmann vdc->get_features = virtio_gpu_get_features; 12379d9e1521SGerd Hoffmann vdc->set_features = virtio_gpu_set_features; 123862232bf4SGerd Hoffmann 123962232bf4SGerd Hoffmann vdc->reset = virtio_gpu_reset; 124062232bf4SGerd Hoffmann 124162232bf4SGerd Hoffmann dc->props = virtio_gpu_properties; 124262232bf4SGerd Hoffmann } 124362232bf4SGerd Hoffmann 124462232bf4SGerd Hoffmann static const TypeInfo virtio_gpu_info = { 124562232bf4SGerd Hoffmann .name = TYPE_VIRTIO_GPU, 124662232bf4SGerd Hoffmann .parent = TYPE_VIRTIO_DEVICE, 124762232bf4SGerd Hoffmann .instance_size = sizeof(VirtIOGPU), 124862232bf4SGerd Hoffmann .instance_init = virtio_gpu_instance_init, 124962232bf4SGerd Hoffmann .class_init = virtio_gpu_class_init, 125062232bf4SGerd Hoffmann }; 125162232bf4SGerd Hoffmann 125262232bf4SGerd Hoffmann static void virtio_register_types(void) 125362232bf4SGerd Hoffmann { 125462232bf4SGerd Hoffmann type_register_static(&virtio_gpu_info); 125562232bf4SGerd Hoffmann } 125662232bf4SGerd Hoffmann 125762232bf4SGerd Hoffmann type_init(virtio_register_types) 125862232bf4SGerd Hoffmann 125962232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctrl_hdr) != 24); 126062232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_update_cursor) != 56); 126162232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_unref) != 32); 126262232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_create_2d) != 40); 126362232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_set_scanout) != 48); 126462232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_flush) != 48); 126562232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_transfer_to_host_2d) != 56); 126662232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_mem_entry) != 16); 126762232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_attach_backing) != 32); 126862232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_detach_backing) != 32); 126962232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_display_info) != 408); 12709d9e1521SGerd Hoffmann 12719d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_transfer_host_3d) != 72); 12729d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_create_3d) != 72); 12739d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_create) != 96); 12749d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_destroy) != 24); 12759d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_resource) != 32); 12769d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_cmd_submit) != 32); 12779d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_get_capset_info) != 32); 12789d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_capset_info) != 40); 12799d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_get_capset) != 32); 12809d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_capset) != 24); 1281