162232bf4SGerd Hoffmann /* 262232bf4SGerd Hoffmann * Virtio GPU Device 362232bf4SGerd Hoffmann * 462232bf4SGerd Hoffmann * Copyright Red Hat, Inc. 2013-2014 562232bf4SGerd Hoffmann * 662232bf4SGerd Hoffmann * Authors: 762232bf4SGerd Hoffmann * Dave Airlie <airlied@redhat.com> 862232bf4SGerd Hoffmann * Gerd Hoffmann <kraxel@redhat.com> 962232bf4SGerd Hoffmann * 102e252145SGerd Hoffmann * This work is licensed under the terms of the GNU GPL, version 2 or later. 1162232bf4SGerd Hoffmann * See the COPYING file in the top-level directory. 1262232bf4SGerd Hoffmann */ 1362232bf4SGerd Hoffmann 1462232bf4SGerd Hoffmann #include "qemu-common.h" 1562232bf4SGerd Hoffmann #include "qemu/iov.h" 1662232bf4SGerd Hoffmann #include "ui/console.h" 1762232bf4SGerd Hoffmann #include "trace.h" 1862232bf4SGerd Hoffmann #include "hw/virtio/virtio.h" 1962232bf4SGerd Hoffmann #include "hw/virtio/virtio-gpu.h" 2062232bf4SGerd Hoffmann #include "hw/virtio/virtio-bus.h" 2162232bf4SGerd Hoffmann 2262232bf4SGerd Hoffmann static struct virtio_gpu_simple_resource* 2362232bf4SGerd Hoffmann virtio_gpu_find_resource(VirtIOGPU *g, uint32_t resource_id); 2462232bf4SGerd Hoffmann 25*9d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 26*9d9e1521SGerd Hoffmann #include "virglrenderer.h" 27*9d9e1521SGerd Hoffmann #define VIRGL(_g, _virgl, _simple, ...) \ 28*9d9e1521SGerd Hoffmann do { \ 29*9d9e1521SGerd Hoffmann if (_g->use_virgl_renderer) { \ 30*9d9e1521SGerd Hoffmann _virgl(__VA_ARGS__); \ 31*9d9e1521SGerd Hoffmann } else { \ 32*9d9e1521SGerd Hoffmann _simple(__VA_ARGS__); \ 33*9d9e1521SGerd Hoffmann } \ 34*9d9e1521SGerd Hoffmann } while (0) 35*9d9e1521SGerd Hoffmann #else 36*9d9e1521SGerd Hoffmann #define VIRGL(_g, _virgl, _simple, ...) \ 37*9d9e1521SGerd Hoffmann do { \ 38*9d9e1521SGerd Hoffmann _simple(__VA_ARGS__); \ 39*9d9e1521SGerd Hoffmann } while (0) 40*9d9e1521SGerd Hoffmann #endif 41*9d9e1521SGerd Hoffmann 4262232bf4SGerd Hoffmann static void update_cursor_data_simple(VirtIOGPU *g, 4362232bf4SGerd Hoffmann struct virtio_gpu_scanout *s, 4462232bf4SGerd Hoffmann uint32_t resource_id) 4562232bf4SGerd Hoffmann { 4662232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 4762232bf4SGerd Hoffmann uint32_t pixels; 4862232bf4SGerd Hoffmann 4962232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, resource_id); 5062232bf4SGerd Hoffmann if (!res) { 5162232bf4SGerd Hoffmann return; 5262232bf4SGerd Hoffmann } 5362232bf4SGerd Hoffmann 5462232bf4SGerd Hoffmann if (pixman_image_get_width(res->image) != s->current_cursor->width || 5562232bf4SGerd Hoffmann pixman_image_get_height(res->image) != s->current_cursor->height) { 5662232bf4SGerd Hoffmann return; 5762232bf4SGerd Hoffmann } 5862232bf4SGerd Hoffmann 5962232bf4SGerd Hoffmann pixels = s->current_cursor->width * s->current_cursor->height; 6062232bf4SGerd Hoffmann memcpy(s->current_cursor->data, 6162232bf4SGerd Hoffmann pixman_image_get_data(res->image), 6262232bf4SGerd Hoffmann pixels * sizeof(uint32_t)); 6362232bf4SGerd Hoffmann } 6462232bf4SGerd Hoffmann 65*9d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 66*9d9e1521SGerd Hoffmann 67*9d9e1521SGerd Hoffmann static void update_cursor_data_virgl(VirtIOGPU *g, 68*9d9e1521SGerd Hoffmann struct virtio_gpu_scanout *s, 69*9d9e1521SGerd Hoffmann uint32_t resource_id) 70*9d9e1521SGerd Hoffmann { 71*9d9e1521SGerd Hoffmann uint32_t width, height; 72*9d9e1521SGerd Hoffmann uint32_t pixels, *data; 73*9d9e1521SGerd Hoffmann 74*9d9e1521SGerd Hoffmann data = virgl_renderer_get_cursor_data(resource_id, &width, &height); 75*9d9e1521SGerd Hoffmann if (!data) { 76*9d9e1521SGerd Hoffmann return; 77*9d9e1521SGerd Hoffmann } 78*9d9e1521SGerd Hoffmann 79*9d9e1521SGerd Hoffmann if (width != s->current_cursor->width || 80*9d9e1521SGerd Hoffmann height != s->current_cursor->height) { 81*9d9e1521SGerd Hoffmann return; 82*9d9e1521SGerd Hoffmann } 83*9d9e1521SGerd Hoffmann 84*9d9e1521SGerd Hoffmann pixels = s->current_cursor->width * s->current_cursor->height; 85*9d9e1521SGerd Hoffmann memcpy(s->current_cursor->data, data, pixels * sizeof(uint32_t)); 86*9d9e1521SGerd Hoffmann free(data); 87*9d9e1521SGerd Hoffmann } 88*9d9e1521SGerd Hoffmann 89*9d9e1521SGerd Hoffmann #endif 90*9d9e1521SGerd Hoffmann 9162232bf4SGerd Hoffmann static void update_cursor(VirtIOGPU *g, struct virtio_gpu_update_cursor *cursor) 9262232bf4SGerd Hoffmann { 9362232bf4SGerd Hoffmann struct virtio_gpu_scanout *s; 9462232bf4SGerd Hoffmann 9562232bf4SGerd Hoffmann if (cursor->pos.scanout_id >= g->conf.max_outputs) { 9662232bf4SGerd Hoffmann return; 9762232bf4SGerd Hoffmann } 9862232bf4SGerd Hoffmann s = &g->scanout[cursor->pos.scanout_id]; 9962232bf4SGerd Hoffmann 10062232bf4SGerd Hoffmann if (cursor->hdr.type != VIRTIO_GPU_CMD_MOVE_CURSOR) { 10162232bf4SGerd Hoffmann if (!s->current_cursor) { 10262232bf4SGerd Hoffmann s->current_cursor = cursor_alloc(64, 64); 10362232bf4SGerd Hoffmann } 10462232bf4SGerd Hoffmann 10562232bf4SGerd Hoffmann s->current_cursor->hot_x = cursor->hot_x; 10662232bf4SGerd Hoffmann s->current_cursor->hot_y = cursor->hot_y; 10762232bf4SGerd Hoffmann 10862232bf4SGerd Hoffmann if (cursor->resource_id > 0) { 109*9d9e1521SGerd Hoffmann VIRGL(g, update_cursor_data_virgl, update_cursor_data_simple, 110*9d9e1521SGerd Hoffmann g, s, cursor->resource_id); 11162232bf4SGerd Hoffmann } 11262232bf4SGerd Hoffmann dpy_cursor_define(s->con, s->current_cursor); 11362232bf4SGerd Hoffmann } 11462232bf4SGerd Hoffmann dpy_mouse_set(s->con, cursor->pos.x, cursor->pos.y, 11562232bf4SGerd Hoffmann cursor->resource_id ? 1 : 0); 11662232bf4SGerd Hoffmann } 11762232bf4SGerd Hoffmann 11862232bf4SGerd Hoffmann static void virtio_gpu_get_config(VirtIODevice *vdev, uint8_t *config) 11962232bf4SGerd Hoffmann { 12062232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 12162232bf4SGerd Hoffmann memcpy(config, &g->virtio_config, sizeof(g->virtio_config)); 12262232bf4SGerd Hoffmann } 12362232bf4SGerd Hoffmann 12462232bf4SGerd Hoffmann static void virtio_gpu_set_config(VirtIODevice *vdev, const uint8_t *config) 12562232bf4SGerd Hoffmann { 12662232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 12762232bf4SGerd Hoffmann struct virtio_gpu_config vgconfig; 12862232bf4SGerd Hoffmann 12962232bf4SGerd Hoffmann memcpy(&vgconfig, config, sizeof(g->virtio_config)); 13062232bf4SGerd Hoffmann 13162232bf4SGerd Hoffmann if (vgconfig.events_clear) { 13262232bf4SGerd Hoffmann g->virtio_config.events_read &= ~vgconfig.events_clear; 13362232bf4SGerd Hoffmann } 13462232bf4SGerd Hoffmann } 13562232bf4SGerd Hoffmann 1369d5b731dSJason Wang static uint64_t virtio_gpu_get_features(VirtIODevice *vdev, uint64_t features, 1379d5b731dSJason Wang Error **errp) 13862232bf4SGerd Hoffmann { 139*9d9e1521SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 140*9d9e1521SGerd Hoffmann 141*9d9e1521SGerd Hoffmann if (virtio_gpu_virgl_enabled(g->conf)) { 142*9d9e1521SGerd Hoffmann features |= (1 << VIRTIO_GPU_FEATURE_VIRGL); 143*9d9e1521SGerd Hoffmann } 14462232bf4SGerd Hoffmann return features; 14562232bf4SGerd Hoffmann } 14662232bf4SGerd Hoffmann 147*9d9e1521SGerd Hoffmann static void virtio_gpu_set_features(VirtIODevice *vdev, uint64_t features) 148*9d9e1521SGerd Hoffmann { 149*9d9e1521SGerd Hoffmann static const uint32_t virgl = (1 << VIRTIO_GPU_FEATURE_VIRGL); 150*9d9e1521SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 151*9d9e1521SGerd Hoffmann 152*9d9e1521SGerd Hoffmann g->use_virgl_renderer = ((features & virgl) == virgl); 153*9d9e1521SGerd Hoffmann trace_virtio_gpu_features(g->use_virgl_renderer); 154*9d9e1521SGerd Hoffmann } 155*9d9e1521SGerd Hoffmann 15662232bf4SGerd Hoffmann static void virtio_gpu_notify_event(VirtIOGPU *g, uint32_t event_type) 15762232bf4SGerd Hoffmann { 15862232bf4SGerd Hoffmann g->virtio_config.events_read |= event_type; 15962232bf4SGerd Hoffmann virtio_notify_config(&g->parent_obj); 16062232bf4SGerd Hoffmann } 16162232bf4SGerd Hoffmann 16262232bf4SGerd Hoffmann static struct virtio_gpu_simple_resource * 16362232bf4SGerd Hoffmann virtio_gpu_find_resource(VirtIOGPU *g, uint32_t resource_id) 16462232bf4SGerd Hoffmann { 16562232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 16662232bf4SGerd Hoffmann 16762232bf4SGerd Hoffmann QTAILQ_FOREACH(res, &g->reslist, next) { 16862232bf4SGerd Hoffmann if (res->resource_id == resource_id) { 16962232bf4SGerd Hoffmann return res; 17062232bf4SGerd Hoffmann } 17162232bf4SGerd Hoffmann } 17262232bf4SGerd Hoffmann return NULL; 17362232bf4SGerd Hoffmann } 17462232bf4SGerd Hoffmann 17562232bf4SGerd Hoffmann void virtio_gpu_ctrl_response(VirtIOGPU *g, 17662232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd, 17762232bf4SGerd Hoffmann struct virtio_gpu_ctrl_hdr *resp, 17862232bf4SGerd Hoffmann size_t resp_len) 17962232bf4SGerd Hoffmann { 18062232bf4SGerd Hoffmann size_t s; 18162232bf4SGerd Hoffmann 18262232bf4SGerd Hoffmann if (cmd->cmd_hdr.flags & VIRTIO_GPU_FLAG_FENCE) { 18362232bf4SGerd Hoffmann resp->flags |= VIRTIO_GPU_FLAG_FENCE; 18462232bf4SGerd Hoffmann resp->fence_id = cmd->cmd_hdr.fence_id; 18562232bf4SGerd Hoffmann resp->ctx_id = cmd->cmd_hdr.ctx_id; 18662232bf4SGerd Hoffmann } 18762232bf4SGerd Hoffmann s = iov_from_buf(cmd->elem.in_sg, cmd->elem.in_num, 0, resp, resp_len); 18862232bf4SGerd Hoffmann if (s != resp_len) { 18962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 19062232bf4SGerd Hoffmann "%s: response size incorrect %zu vs %zu\n", 19162232bf4SGerd Hoffmann __func__, s, resp_len); 19262232bf4SGerd Hoffmann } 19362232bf4SGerd Hoffmann virtqueue_push(cmd->vq, &cmd->elem, s); 19462232bf4SGerd Hoffmann virtio_notify(VIRTIO_DEVICE(g), cmd->vq); 19562232bf4SGerd Hoffmann cmd->finished = true; 19662232bf4SGerd Hoffmann } 19762232bf4SGerd Hoffmann 19862232bf4SGerd Hoffmann void virtio_gpu_ctrl_response_nodata(VirtIOGPU *g, 19962232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd, 20062232bf4SGerd Hoffmann enum virtio_gpu_ctrl_type type) 20162232bf4SGerd Hoffmann { 20262232bf4SGerd Hoffmann struct virtio_gpu_ctrl_hdr resp; 20362232bf4SGerd Hoffmann 20462232bf4SGerd Hoffmann memset(&resp, 0, sizeof(resp)); 20562232bf4SGerd Hoffmann resp.type = type; 20662232bf4SGerd Hoffmann virtio_gpu_ctrl_response(g, cmd, &resp, sizeof(resp)); 20762232bf4SGerd Hoffmann } 20862232bf4SGerd Hoffmann 20962232bf4SGerd Hoffmann static void 21062232bf4SGerd Hoffmann virtio_gpu_fill_display_info(VirtIOGPU *g, 21162232bf4SGerd Hoffmann struct virtio_gpu_resp_display_info *dpy_info) 21262232bf4SGerd Hoffmann { 21362232bf4SGerd Hoffmann int i; 21462232bf4SGerd Hoffmann 21562232bf4SGerd Hoffmann for (i = 0; i < g->conf.max_outputs; i++) { 21662232bf4SGerd Hoffmann if (g->enabled_output_bitmask & (1 << i)) { 21762232bf4SGerd Hoffmann dpy_info->pmodes[i].enabled = 1; 21862232bf4SGerd Hoffmann dpy_info->pmodes[i].r.width = g->req_state[i].width; 21962232bf4SGerd Hoffmann dpy_info->pmodes[i].r.height = g->req_state[i].height; 22062232bf4SGerd Hoffmann } 22162232bf4SGerd Hoffmann } 22262232bf4SGerd Hoffmann } 22362232bf4SGerd Hoffmann 22462232bf4SGerd Hoffmann void virtio_gpu_get_display_info(VirtIOGPU *g, 22562232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 22662232bf4SGerd Hoffmann { 22762232bf4SGerd Hoffmann struct virtio_gpu_resp_display_info display_info; 22862232bf4SGerd Hoffmann 22962232bf4SGerd Hoffmann trace_virtio_gpu_cmd_get_display_info(); 23062232bf4SGerd Hoffmann memset(&display_info, 0, sizeof(display_info)); 23162232bf4SGerd Hoffmann display_info.hdr.type = VIRTIO_GPU_RESP_OK_DISPLAY_INFO; 23262232bf4SGerd Hoffmann virtio_gpu_fill_display_info(g, &display_info); 23362232bf4SGerd Hoffmann virtio_gpu_ctrl_response(g, cmd, &display_info.hdr, 23462232bf4SGerd Hoffmann sizeof(display_info)); 23562232bf4SGerd Hoffmann } 23662232bf4SGerd Hoffmann 23762232bf4SGerd Hoffmann static pixman_format_code_t get_pixman_format(uint32_t virtio_gpu_format) 23862232bf4SGerd Hoffmann { 23962232bf4SGerd Hoffmann switch (virtio_gpu_format) { 24062232bf4SGerd Hoffmann #ifdef HOST_WORDS_BIGENDIAN 24162232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8X8_UNORM: 24262232bf4SGerd Hoffmann return PIXMAN_b8g8r8x8; 24362232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8A8_UNORM: 24462232bf4SGerd Hoffmann return PIXMAN_b8g8r8a8; 24562232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8R8G8B8_UNORM: 24662232bf4SGerd Hoffmann return PIXMAN_x8r8g8b8; 24762232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8R8G8B8_UNORM: 24862232bf4SGerd Hoffmann return PIXMAN_a8r8g8b8; 24962232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8X8_UNORM: 25062232bf4SGerd Hoffmann return PIXMAN_r8g8b8x8; 25162232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8A8_UNORM: 25262232bf4SGerd Hoffmann return PIXMAN_r8g8b8a8; 25362232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8B8G8R8_UNORM: 25462232bf4SGerd Hoffmann return PIXMAN_x8b8g8r8; 25562232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8B8G8R8_UNORM: 25662232bf4SGerd Hoffmann return PIXMAN_a8b8g8r8; 25762232bf4SGerd Hoffmann #else 25862232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8X8_UNORM: 25962232bf4SGerd Hoffmann return PIXMAN_x8r8g8b8; 26062232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_B8G8R8A8_UNORM: 26162232bf4SGerd Hoffmann return PIXMAN_a8r8g8b8; 26262232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8R8G8B8_UNORM: 26362232bf4SGerd Hoffmann return PIXMAN_b8g8r8x8; 26462232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8R8G8B8_UNORM: 26562232bf4SGerd Hoffmann return PIXMAN_b8g8r8a8; 26662232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8X8_UNORM: 26762232bf4SGerd Hoffmann return PIXMAN_x8b8g8r8; 26862232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_R8G8B8A8_UNORM: 26962232bf4SGerd Hoffmann return PIXMAN_a8b8g8r8; 27062232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_X8B8G8R8_UNORM: 27162232bf4SGerd Hoffmann return PIXMAN_r8g8b8x8; 27262232bf4SGerd Hoffmann case VIRTIO_GPU_FORMAT_A8B8G8R8_UNORM: 27362232bf4SGerd Hoffmann return PIXMAN_r8g8b8a8; 27462232bf4SGerd Hoffmann #endif 27562232bf4SGerd Hoffmann default: 27662232bf4SGerd Hoffmann return 0; 27762232bf4SGerd Hoffmann } 27862232bf4SGerd Hoffmann } 27962232bf4SGerd Hoffmann 28062232bf4SGerd Hoffmann static void virtio_gpu_resource_create_2d(VirtIOGPU *g, 28162232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 28262232bf4SGerd Hoffmann { 28362232bf4SGerd Hoffmann pixman_format_code_t pformat; 28462232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 28562232bf4SGerd Hoffmann struct virtio_gpu_resource_create_2d c2d; 28662232bf4SGerd Hoffmann 28762232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(c2d); 28862232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_create_2d(c2d.resource_id, c2d.format, 28962232bf4SGerd Hoffmann c2d.width, c2d.height); 29062232bf4SGerd Hoffmann 29162232bf4SGerd Hoffmann if (c2d.resource_id == 0) { 29262232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: resource id 0 is not allowed\n", 29362232bf4SGerd Hoffmann __func__); 29462232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 29562232bf4SGerd Hoffmann return; 29662232bf4SGerd Hoffmann } 29762232bf4SGerd Hoffmann 29862232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, c2d.resource_id); 29962232bf4SGerd Hoffmann if (res) { 30062232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: resource already exists %d\n", 30162232bf4SGerd Hoffmann __func__, c2d.resource_id); 30262232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 30362232bf4SGerd Hoffmann return; 30462232bf4SGerd Hoffmann } 30562232bf4SGerd Hoffmann 30662232bf4SGerd Hoffmann res = g_new0(struct virtio_gpu_simple_resource, 1); 30762232bf4SGerd Hoffmann 30862232bf4SGerd Hoffmann res->width = c2d.width; 30962232bf4SGerd Hoffmann res->height = c2d.height; 31062232bf4SGerd Hoffmann res->format = c2d.format; 31162232bf4SGerd Hoffmann res->resource_id = c2d.resource_id; 31262232bf4SGerd Hoffmann 31362232bf4SGerd Hoffmann pformat = get_pixman_format(c2d.format); 31462232bf4SGerd Hoffmann if (!pformat) { 31562232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 31662232bf4SGerd Hoffmann "%s: host couldn't handle guest format %d\n", 31762232bf4SGerd Hoffmann __func__, c2d.format); 31862232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 31962232bf4SGerd Hoffmann return; 32062232bf4SGerd Hoffmann } 32162232bf4SGerd Hoffmann res->image = pixman_image_create_bits(pformat, 32262232bf4SGerd Hoffmann c2d.width, 32362232bf4SGerd Hoffmann c2d.height, 32462232bf4SGerd Hoffmann NULL, 0); 32562232bf4SGerd Hoffmann 32662232bf4SGerd Hoffmann if (!res->image) { 32762232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 32862232bf4SGerd Hoffmann "%s: resource creation failed %d %d %d\n", 32962232bf4SGerd Hoffmann __func__, c2d.resource_id, c2d.width, c2d.height); 33062232bf4SGerd Hoffmann g_free(res); 33162232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_OUT_OF_MEMORY; 33262232bf4SGerd Hoffmann return; 33362232bf4SGerd Hoffmann } 33462232bf4SGerd Hoffmann 33562232bf4SGerd Hoffmann QTAILQ_INSERT_HEAD(&g->reslist, res, next); 33662232bf4SGerd Hoffmann } 33762232bf4SGerd Hoffmann 33862232bf4SGerd Hoffmann static void virtio_gpu_resource_destroy(VirtIOGPU *g, 33962232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res) 34062232bf4SGerd Hoffmann { 34162232bf4SGerd Hoffmann pixman_image_unref(res->image); 34262232bf4SGerd Hoffmann QTAILQ_REMOVE(&g->reslist, res, next); 34362232bf4SGerd Hoffmann g_free(res); 34462232bf4SGerd Hoffmann } 34562232bf4SGerd Hoffmann 34662232bf4SGerd Hoffmann static void virtio_gpu_resource_unref(VirtIOGPU *g, 34762232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 34862232bf4SGerd Hoffmann { 34962232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 35062232bf4SGerd Hoffmann struct virtio_gpu_resource_unref unref; 35162232bf4SGerd Hoffmann 35262232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(unref); 35362232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_unref(unref.resource_id); 35462232bf4SGerd Hoffmann 35562232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, unref.resource_id); 35662232bf4SGerd Hoffmann if (!res) { 35762232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 35862232bf4SGerd Hoffmann __func__, unref.resource_id); 35962232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 36062232bf4SGerd Hoffmann return; 36162232bf4SGerd Hoffmann } 36262232bf4SGerd Hoffmann virtio_gpu_resource_destroy(g, res); 36362232bf4SGerd Hoffmann } 36462232bf4SGerd Hoffmann 36562232bf4SGerd Hoffmann static void virtio_gpu_transfer_to_host_2d(VirtIOGPU *g, 36662232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 36762232bf4SGerd Hoffmann { 36862232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 36962232bf4SGerd Hoffmann int h; 37062232bf4SGerd Hoffmann uint32_t src_offset, dst_offset, stride; 37162232bf4SGerd Hoffmann int bpp; 37262232bf4SGerd Hoffmann pixman_format_code_t format; 37362232bf4SGerd Hoffmann struct virtio_gpu_transfer_to_host_2d t2d; 37462232bf4SGerd Hoffmann 37562232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(t2d); 37662232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_xfer_toh_2d(t2d.resource_id); 37762232bf4SGerd Hoffmann 37862232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, t2d.resource_id); 37962232bf4SGerd Hoffmann if (!res || !res->iov) { 38062232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 38162232bf4SGerd Hoffmann __func__, t2d.resource_id); 38262232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 38362232bf4SGerd Hoffmann return; 38462232bf4SGerd Hoffmann } 38562232bf4SGerd Hoffmann 38662232bf4SGerd Hoffmann if (t2d.r.x > res->width || 38762232bf4SGerd Hoffmann t2d.r.y > res->height || 38862232bf4SGerd Hoffmann t2d.r.width > res->width || 38962232bf4SGerd Hoffmann t2d.r.height > res->height || 39062232bf4SGerd Hoffmann t2d.r.x + t2d.r.width > res->width || 39162232bf4SGerd Hoffmann t2d.r.y + t2d.r.height > res->height) { 39262232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: transfer bounds outside resource" 39362232bf4SGerd Hoffmann " bounds for resource %d: %d %d %d %d vs %d %d\n", 39462232bf4SGerd Hoffmann __func__, t2d.resource_id, t2d.r.x, t2d.r.y, 39562232bf4SGerd Hoffmann t2d.r.width, t2d.r.height, res->width, res->height); 39662232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 39762232bf4SGerd Hoffmann return; 39862232bf4SGerd Hoffmann } 39962232bf4SGerd Hoffmann 40062232bf4SGerd Hoffmann format = pixman_image_get_format(res->image); 40162232bf4SGerd Hoffmann bpp = (PIXMAN_FORMAT_BPP(format) + 7) / 8; 40262232bf4SGerd Hoffmann stride = pixman_image_get_stride(res->image); 40362232bf4SGerd Hoffmann 40462232bf4SGerd Hoffmann if (t2d.offset || t2d.r.x || t2d.r.y || 40562232bf4SGerd Hoffmann t2d.r.width != pixman_image_get_width(res->image)) { 40662232bf4SGerd Hoffmann void *img_data = pixman_image_get_data(res->image); 40762232bf4SGerd Hoffmann for (h = 0; h < t2d.r.height; h++) { 40862232bf4SGerd Hoffmann src_offset = t2d.offset + stride * h; 40962232bf4SGerd Hoffmann dst_offset = (t2d.r.y + h) * stride + (t2d.r.x * bpp); 41062232bf4SGerd Hoffmann 41162232bf4SGerd Hoffmann iov_to_buf(res->iov, res->iov_cnt, src_offset, 41262232bf4SGerd Hoffmann (uint8_t *)img_data 41362232bf4SGerd Hoffmann + dst_offset, t2d.r.width * bpp); 41462232bf4SGerd Hoffmann } 41562232bf4SGerd Hoffmann } else { 41662232bf4SGerd Hoffmann iov_to_buf(res->iov, res->iov_cnt, 0, 41762232bf4SGerd Hoffmann pixman_image_get_data(res->image), 41862232bf4SGerd Hoffmann pixman_image_get_stride(res->image) 41962232bf4SGerd Hoffmann * pixman_image_get_height(res->image)); 42062232bf4SGerd Hoffmann } 42162232bf4SGerd Hoffmann } 42262232bf4SGerd Hoffmann 42362232bf4SGerd Hoffmann static void virtio_gpu_resource_flush(VirtIOGPU *g, 42462232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 42562232bf4SGerd Hoffmann { 42662232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 42762232bf4SGerd Hoffmann struct virtio_gpu_resource_flush rf; 42862232bf4SGerd Hoffmann pixman_region16_t flush_region; 42962232bf4SGerd Hoffmann int i; 43062232bf4SGerd Hoffmann 43162232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(rf); 43262232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_flush(rf.resource_id, 43362232bf4SGerd Hoffmann rf.r.width, rf.r.height, rf.r.x, rf.r.y); 43462232bf4SGerd Hoffmann 43562232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, rf.resource_id); 43662232bf4SGerd Hoffmann if (!res) { 43762232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 43862232bf4SGerd Hoffmann __func__, rf.resource_id); 43962232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 44062232bf4SGerd Hoffmann return; 44162232bf4SGerd Hoffmann } 44262232bf4SGerd Hoffmann 44362232bf4SGerd Hoffmann if (rf.r.x > res->width || 44462232bf4SGerd Hoffmann rf.r.y > res->height || 44562232bf4SGerd Hoffmann rf.r.width > res->width || 44662232bf4SGerd Hoffmann rf.r.height > res->height || 44762232bf4SGerd Hoffmann rf.r.x + rf.r.width > res->width || 44862232bf4SGerd Hoffmann rf.r.y + rf.r.height > res->height) { 44962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: flush bounds outside resource" 45062232bf4SGerd Hoffmann " bounds for resource %d: %d %d %d %d vs %d %d\n", 45162232bf4SGerd Hoffmann __func__, rf.resource_id, rf.r.x, rf.r.y, 45262232bf4SGerd Hoffmann rf.r.width, rf.r.height, res->width, res->height); 45362232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 45462232bf4SGerd Hoffmann return; 45562232bf4SGerd Hoffmann } 45662232bf4SGerd Hoffmann 45762232bf4SGerd Hoffmann pixman_region_init_rect(&flush_region, 45862232bf4SGerd Hoffmann rf.r.x, rf.r.y, rf.r.width, rf.r.height); 45962232bf4SGerd Hoffmann for (i = 0; i < VIRTIO_GPU_MAX_SCANOUT; i++) { 46062232bf4SGerd Hoffmann struct virtio_gpu_scanout *scanout; 46162232bf4SGerd Hoffmann pixman_region16_t region, finalregion; 46262232bf4SGerd Hoffmann pixman_box16_t *extents; 46362232bf4SGerd Hoffmann 46462232bf4SGerd Hoffmann if (!(res->scanout_bitmask & (1 << i))) { 46562232bf4SGerd Hoffmann continue; 46662232bf4SGerd Hoffmann } 46762232bf4SGerd Hoffmann scanout = &g->scanout[i]; 46862232bf4SGerd Hoffmann 46962232bf4SGerd Hoffmann pixman_region_init(&finalregion); 47062232bf4SGerd Hoffmann pixman_region_init_rect(®ion, scanout->x, scanout->y, 47162232bf4SGerd Hoffmann scanout->width, scanout->height); 47262232bf4SGerd Hoffmann 47362232bf4SGerd Hoffmann pixman_region_intersect(&finalregion, &flush_region, ®ion); 47462232bf4SGerd Hoffmann pixman_region_translate(&finalregion, -scanout->x, -scanout->y); 47562232bf4SGerd Hoffmann extents = pixman_region_extents(&finalregion); 47662232bf4SGerd Hoffmann /* work out the area we need to update for each console */ 47762232bf4SGerd Hoffmann dpy_gfx_update(g->scanout[i].con, 47862232bf4SGerd Hoffmann extents->x1, extents->y1, 47962232bf4SGerd Hoffmann extents->x2 - extents->x1, 48062232bf4SGerd Hoffmann extents->y2 - extents->y1); 48162232bf4SGerd Hoffmann 48262232bf4SGerd Hoffmann pixman_region_fini(®ion); 48362232bf4SGerd Hoffmann pixman_region_fini(&finalregion); 48462232bf4SGerd Hoffmann } 48562232bf4SGerd Hoffmann pixman_region_fini(&flush_region); 48662232bf4SGerd Hoffmann } 48762232bf4SGerd Hoffmann 48862232bf4SGerd Hoffmann static void virtio_gpu_set_scanout(VirtIOGPU *g, 48962232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 49062232bf4SGerd Hoffmann { 49162232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 49262232bf4SGerd Hoffmann struct virtio_gpu_scanout *scanout; 49362232bf4SGerd Hoffmann pixman_format_code_t format; 49462232bf4SGerd Hoffmann uint32_t offset; 49562232bf4SGerd Hoffmann int bpp; 49662232bf4SGerd Hoffmann struct virtio_gpu_set_scanout ss; 49762232bf4SGerd Hoffmann 49862232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(ss); 49962232bf4SGerd Hoffmann trace_virtio_gpu_cmd_set_scanout(ss.scanout_id, ss.resource_id, 50062232bf4SGerd Hoffmann ss.r.width, ss.r.height, ss.r.x, ss.r.y); 50162232bf4SGerd Hoffmann 50262232bf4SGerd Hoffmann g->enable = 1; 50362232bf4SGerd Hoffmann if (ss.resource_id == 0) { 50462232bf4SGerd Hoffmann scanout = &g->scanout[ss.scanout_id]; 50562232bf4SGerd Hoffmann if (scanout->resource_id) { 50662232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, scanout->resource_id); 50762232bf4SGerd Hoffmann if (res) { 50862232bf4SGerd Hoffmann res->scanout_bitmask &= ~(1 << ss.scanout_id); 50962232bf4SGerd Hoffmann } 51062232bf4SGerd Hoffmann } 51162232bf4SGerd Hoffmann if (ss.scanout_id == 0 || 51262232bf4SGerd Hoffmann ss.scanout_id >= g->conf.max_outputs) { 51362232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 51462232bf4SGerd Hoffmann "%s: illegal scanout id specified %d", 51562232bf4SGerd Hoffmann __func__, ss.scanout_id); 51662232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_SCANOUT_ID; 51762232bf4SGerd Hoffmann return; 51862232bf4SGerd Hoffmann } 51962232bf4SGerd Hoffmann dpy_gfx_replace_surface(g->scanout[ss.scanout_id].con, NULL); 52062232bf4SGerd Hoffmann scanout->ds = NULL; 52162232bf4SGerd Hoffmann scanout->width = 0; 52262232bf4SGerd Hoffmann scanout->height = 0; 52362232bf4SGerd Hoffmann return; 52462232bf4SGerd Hoffmann } 52562232bf4SGerd Hoffmann 52662232bf4SGerd Hoffmann /* create a surface for this scanout */ 52762232bf4SGerd Hoffmann if (ss.scanout_id >= VIRTIO_GPU_MAX_SCANOUT || 52862232bf4SGerd Hoffmann ss.scanout_id >= g->conf.max_outputs) { 52962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout id specified %d", 53062232bf4SGerd Hoffmann __func__, ss.scanout_id); 53162232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_SCANOUT_ID; 53262232bf4SGerd Hoffmann return; 53362232bf4SGerd Hoffmann } 53462232bf4SGerd Hoffmann 53562232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, ss.resource_id); 53662232bf4SGerd Hoffmann if (!res) { 53762232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 53862232bf4SGerd Hoffmann __func__, ss.resource_id); 53962232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 54062232bf4SGerd Hoffmann return; 54162232bf4SGerd Hoffmann } 54262232bf4SGerd Hoffmann 54362232bf4SGerd Hoffmann if (ss.r.x > res->width || 54462232bf4SGerd Hoffmann ss.r.y > res->height || 54562232bf4SGerd Hoffmann ss.r.width > res->width || 54662232bf4SGerd Hoffmann ss.r.height > res->height || 54762232bf4SGerd Hoffmann ss.r.x + ss.r.width > res->width || 54862232bf4SGerd Hoffmann ss.r.y + ss.r.height > res->height) { 54962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal scanout %d bounds for" 55062232bf4SGerd Hoffmann " resource %d, (%d,%d)+%d,%d vs %d %d\n", 55162232bf4SGerd Hoffmann __func__, ss.scanout_id, ss.resource_id, ss.r.x, ss.r.y, 55262232bf4SGerd Hoffmann ss.r.width, ss.r.height, res->width, res->height); 55362232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_PARAMETER; 55462232bf4SGerd Hoffmann return; 55562232bf4SGerd Hoffmann } 55662232bf4SGerd Hoffmann 55762232bf4SGerd Hoffmann scanout = &g->scanout[ss.scanout_id]; 55862232bf4SGerd Hoffmann 55962232bf4SGerd Hoffmann format = pixman_image_get_format(res->image); 56062232bf4SGerd Hoffmann bpp = (PIXMAN_FORMAT_BPP(format) + 7) / 8; 56162232bf4SGerd Hoffmann offset = (ss.r.x * bpp) + ss.r.y * pixman_image_get_stride(res->image); 56262232bf4SGerd Hoffmann if (!scanout->ds || surface_data(scanout->ds) 56362232bf4SGerd Hoffmann != ((uint8_t *)pixman_image_get_data(res->image) + offset) || 56462232bf4SGerd Hoffmann scanout->width != ss.r.width || 56562232bf4SGerd Hoffmann scanout->height != ss.r.height) { 56662232bf4SGerd Hoffmann /* realloc the surface ptr */ 56762232bf4SGerd Hoffmann scanout->ds = qemu_create_displaysurface_from 56862232bf4SGerd Hoffmann (ss.r.width, ss.r.height, format, 56962232bf4SGerd Hoffmann pixman_image_get_stride(res->image), 57062232bf4SGerd Hoffmann (uint8_t *)pixman_image_get_data(res->image) + offset); 57162232bf4SGerd Hoffmann if (!scanout->ds) { 57262232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 57362232bf4SGerd Hoffmann return; 57462232bf4SGerd Hoffmann } 57562232bf4SGerd Hoffmann dpy_gfx_replace_surface(g->scanout[ss.scanout_id].con, scanout->ds); 57662232bf4SGerd Hoffmann } 57762232bf4SGerd Hoffmann 57862232bf4SGerd Hoffmann res->scanout_bitmask |= (1 << ss.scanout_id); 57962232bf4SGerd Hoffmann scanout->resource_id = ss.resource_id; 58062232bf4SGerd Hoffmann scanout->x = ss.r.x; 58162232bf4SGerd Hoffmann scanout->y = ss.r.y; 58262232bf4SGerd Hoffmann scanout->width = ss.r.width; 58362232bf4SGerd Hoffmann scanout->height = ss.r.height; 58462232bf4SGerd Hoffmann } 58562232bf4SGerd Hoffmann 58662232bf4SGerd Hoffmann int virtio_gpu_create_mapping_iov(struct virtio_gpu_resource_attach_backing *ab, 58762232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd, 58862232bf4SGerd Hoffmann struct iovec **iov) 58962232bf4SGerd Hoffmann { 59062232bf4SGerd Hoffmann struct virtio_gpu_mem_entry *ents; 59162232bf4SGerd Hoffmann size_t esize, s; 59262232bf4SGerd Hoffmann int i; 59362232bf4SGerd Hoffmann 59462232bf4SGerd Hoffmann if (ab->nr_entries > 16384) { 59562232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 5962c84167bSGerd Hoffmann "%s: nr_entries is too big (%d > 16384)\n", 59762232bf4SGerd Hoffmann __func__, ab->nr_entries); 59862232bf4SGerd Hoffmann return -1; 59962232bf4SGerd Hoffmann } 60062232bf4SGerd Hoffmann 60162232bf4SGerd Hoffmann esize = sizeof(*ents) * ab->nr_entries; 60262232bf4SGerd Hoffmann ents = g_malloc(esize); 60362232bf4SGerd Hoffmann s = iov_to_buf(cmd->elem.out_sg, cmd->elem.out_num, 60462232bf4SGerd Hoffmann sizeof(*ab), ents, esize); 60562232bf4SGerd Hoffmann if (s != esize) { 60662232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 60762232bf4SGerd Hoffmann "%s: command data size incorrect %zu vs %zu\n", 60862232bf4SGerd Hoffmann __func__, s, esize); 60962232bf4SGerd Hoffmann g_free(ents); 61062232bf4SGerd Hoffmann return -1; 61162232bf4SGerd Hoffmann } 61262232bf4SGerd Hoffmann 61362232bf4SGerd Hoffmann *iov = g_malloc0(sizeof(struct iovec) * ab->nr_entries); 61462232bf4SGerd Hoffmann for (i = 0; i < ab->nr_entries; i++) { 61562232bf4SGerd Hoffmann hwaddr len = ents[i].length; 61662232bf4SGerd Hoffmann (*iov)[i].iov_len = ents[i].length; 61762232bf4SGerd Hoffmann (*iov)[i].iov_base = cpu_physical_memory_map(ents[i].addr, &len, 1); 61862232bf4SGerd Hoffmann if (!(*iov)[i].iov_base || len != ents[i].length) { 61962232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: failed to map MMIO memory for" 62062232bf4SGerd Hoffmann " resource %d element %d\n", 62162232bf4SGerd Hoffmann __func__, ab->resource_id, i); 62262232bf4SGerd Hoffmann virtio_gpu_cleanup_mapping_iov(*iov, i); 62362232bf4SGerd Hoffmann g_free(ents); 62462232bf4SGerd Hoffmann *iov = NULL; 62562232bf4SGerd Hoffmann return -1; 62662232bf4SGerd Hoffmann } 62762232bf4SGerd Hoffmann } 62862232bf4SGerd Hoffmann g_free(ents); 62962232bf4SGerd Hoffmann return 0; 63062232bf4SGerd Hoffmann } 63162232bf4SGerd Hoffmann 63262232bf4SGerd Hoffmann void virtio_gpu_cleanup_mapping_iov(struct iovec *iov, uint32_t count) 63362232bf4SGerd Hoffmann { 63462232bf4SGerd Hoffmann int i; 63562232bf4SGerd Hoffmann 63662232bf4SGerd Hoffmann for (i = 0; i < count; i++) { 63762232bf4SGerd Hoffmann cpu_physical_memory_unmap(iov[i].iov_base, iov[i].iov_len, 1, 63862232bf4SGerd Hoffmann iov[i].iov_len); 63962232bf4SGerd Hoffmann } 6407f3be0f2SGerd Hoffmann g_free(iov); 64162232bf4SGerd Hoffmann } 64262232bf4SGerd Hoffmann 64362232bf4SGerd Hoffmann static void virtio_gpu_cleanup_mapping(struct virtio_gpu_simple_resource *res) 64462232bf4SGerd Hoffmann { 64562232bf4SGerd Hoffmann virtio_gpu_cleanup_mapping_iov(res->iov, res->iov_cnt); 64662232bf4SGerd Hoffmann res->iov = NULL; 64762232bf4SGerd Hoffmann res->iov_cnt = 0; 64862232bf4SGerd Hoffmann } 64962232bf4SGerd Hoffmann 65062232bf4SGerd Hoffmann static void 65162232bf4SGerd Hoffmann virtio_gpu_resource_attach_backing(VirtIOGPU *g, 65262232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 65362232bf4SGerd Hoffmann { 65462232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 65562232bf4SGerd Hoffmann struct virtio_gpu_resource_attach_backing ab; 65662232bf4SGerd Hoffmann int ret; 65762232bf4SGerd Hoffmann 65862232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(ab); 65962232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_back_attach(ab.resource_id); 66062232bf4SGerd Hoffmann 66162232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, ab.resource_id); 66262232bf4SGerd Hoffmann if (!res) { 66362232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 66462232bf4SGerd Hoffmann __func__, ab.resource_id); 66562232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 66662232bf4SGerd Hoffmann return; 66762232bf4SGerd Hoffmann } 66862232bf4SGerd Hoffmann 66962232bf4SGerd Hoffmann ret = virtio_gpu_create_mapping_iov(&ab, cmd, &res->iov); 67062232bf4SGerd Hoffmann if (ret != 0) { 67162232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 67262232bf4SGerd Hoffmann return; 67362232bf4SGerd Hoffmann } 67462232bf4SGerd Hoffmann 67562232bf4SGerd Hoffmann res->iov_cnt = ab.nr_entries; 67662232bf4SGerd Hoffmann } 67762232bf4SGerd Hoffmann 67862232bf4SGerd Hoffmann static void 67962232bf4SGerd Hoffmann virtio_gpu_resource_detach_backing(VirtIOGPU *g, 68062232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 68162232bf4SGerd Hoffmann { 68262232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res; 68362232bf4SGerd Hoffmann struct virtio_gpu_resource_detach_backing detach; 68462232bf4SGerd Hoffmann 68562232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(detach); 68662232bf4SGerd Hoffmann trace_virtio_gpu_cmd_res_back_detach(detach.resource_id); 68762232bf4SGerd Hoffmann 68862232bf4SGerd Hoffmann res = virtio_gpu_find_resource(g, detach.resource_id); 68962232bf4SGerd Hoffmann if (!res || !res->iov) { 69062232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, "%s: illegal resource specified %d\n", 69162232bf4SGerd Hoffmann __func__, detach.resource_id); 69262232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_INVALID_RESOURCE_ID; 69362232bf4SGerd Hoffmann return; 69462232bf4SGerd Hoffmann } 69562232bf4SGerd Hoffmann virtio_gpu_cleanup_mapping(res); 69662232bf4SGerd Hoffmann } 69762232bf4SGerd Hoffmann 69862232bf4SGerd Hoffmann static void virtio_gpu_simple_process_cmd(VirtIOGPU *g, 69962232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd) 70062232bf4SGerd Hoffmann { 70162232bf4SGerd Hoffmann VIRTIO_GPU_FILL_CMD(cmd->cmd_hdr); 70262232bf4SGerd Hoffmann 70362232bf4SGerd Hoffmann switch (cmd->cmd_hdr.type) { 70462232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_GET_DISPLAY_INFO: 70562232bf4SGerd Hoffmann virtio_gpu_get_display_info(g, cmd); 70662232bf4SGerd Hoffmann break; 70762232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_CREATE_2D: 70862232bf4SGerd Hoffmann virtio_gpu_resource_create_2d(g, cmd); 70962232bf4SGerd Hoffmann break; 71062232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_UNREF: 71162232bf4SGerd Hoffmann virtio_gpu_resource_unref(g, cmd); 71262232bf4SGerd Hoffmann break; 71362232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_FLUSH: 71462232bf4SGerd Hoffmann virtio_gpu_resource_flush(g, cmd); 71562232bf4SGerd Hoffmann break; 71662232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_TRANSFER_TO_HOST_2D: 71762232bf4SGerd Hoffmann virtio_gpu_transfer_to_host_2d(g, cmd); 71862232bf4SGerd Hoffmann break; 71962232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_SET_SCANOUT: 72062232bf4SGerd Hoffmann virtio_gpu_set_scanout(g, cmd); 72162232bf4SGerd Hoffmann break; 72262232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_ATTACH_BACKING: 72362232bf4SGerd Hoffmann virtio_gpu_resource_attach_backing(g, cmd); 72462232bf4SGerd Hoffmann break; 72562232bf4SGerd Hoffmann case VIRTIO_GPU_CMD_RESOURCE_DETACH_BACKING: 72662232bf4SGerd Hoffmann virtio_gpu_resource_detach_backing(g, cmd); 72762232bf4SGerd Hoffmann break; 72862232bf4SGerd Hoffmann default: 72962232bf4SGerd Hoffmann cmd->error = VIRTIO_GPU_RESP_ERR_UNSPEC; 73062232bf4SGerd Hoffmann break; 73162232bf4SGerd Hoffmann } 73262232bf4SGerd Hoffmann if (!cmd->finished) { 73362232bf4SGerd Hoffmann virtio_gpu_ctrl_response_nodata(g, cmd, cmd->error ? cmd->error : 73462232bf4SGerd Hoffmann VIRTIO_GPU_RESP_OK_NODATA); 73562232bf4SGerd Hoffmann } 73662232bf4SGerd Hoffmann } 73762232bf4SGerd Hoffmann 73862232bf4SGerd Hoffmann static void virtio_gpu_handle_ctrl_cb(VirtIODevice *vdev, VirtQueue *vq) 73962232bf4SGerd Hoffmann { 74062232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 74162232bf4SGerd Hoffmann qemu_bh_schedule(g->ctrl_bh); 74262232bf4SGerd Hoffmann } 74362232bf4SGerd Hoffmann 74462232bf4SGerd Hoffmann static void virtio_gpu_handle_cursor_cb(VirtIODevice *vdev, VirtQueue *vq) 74562232bf4SGerd Hoffmann { 74662232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 74762232bf4SGerd Hoffmann qemu_bh_schedule(g->cursor_bh); 74862232bf4SGerd Hoffmann } 74962232bf4SGerd Hoffmann 75062232bf4SGerd Hoffmann static void virtio_gpu_handle_ctrl(VirtIODevice *vdev, VirtQueue *vq) 75162232bf4SGerd Hoffmann { 75262232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 75362232bf4SGerd Hoffmann struct virtio_gpu_ctrl_command *cmd; 75462232bf4SGerd Hoffmann 75562232bf4SGerd Hoffmann if (!virtio_queue_ready(vq)) { 75662232bf4SGerd Hoffmann return; 75762232bf4SGerd Hoffmann } 75862232bf4SGerd Hoffmann 759*9d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 760*9d9e1521SGerd Hoffmann if (!g->renderer_inited && g->use_virgl_renderer) { 761*9d9e1521SGerd Hoffmann virtio_gpu_virgl_init(g); 762*9d9e1521SGerd Hoffmann g->renderer_inited = true; 763*9d9e1521SGerd Hoffmann } 764*9d9e1521SGerd Hoffmann #endif 765*9d9e1521SGerd Hoffmann 76662232bf4SGerd Hoffmann cmd = g_new(struct virtio_gpu_ctrl_command, 1); 76762232bf4SGerd Hoffmann while (virtqueue_pop(vq, &cmd->elem)) { 76862232bf4SGerd Hoffmann cmd->vq = vq; 76962232bf4SGerd Hoffmann cmd->error = 0; 77062232bf4SGerd Hoffmann cmd->finished = false; 771*9d9e1521SGerd Hoffmann if (virtio_gpu_stats_enabled(g->conf)) { 77262232bf4SGerd Hoffmann g->stats.requests++; 773*9d9e1521SGerd Hoffmann } 77462232bf4SGerd Hoffmann 775*9d9e1521SGerd Hoffmann VIRGL(g, virtio_gpu_virgl_process_cmd, virtio_gpu_simple_process_cmd, 776*9d9e1521SGerd Hoffmann g, cmd); 77762232bf4SGerd Hoffmann if (!cmd->finished) { 77862232bf4SGerd Hoffmann QTAILQ_INSERT_TAIL(&g->fenceq, cmd, next); 779*9d9e1521SGerd Hoffmann g->inflight++; 780*9d9e1521SGerd Hoffmann if (virtio_gpu_stats_enabled(g->conf)) { 781*9d9e1521SGerd Hoffmann if (g->stats.max_inflight < g->inflight) { 782*9d9e1521SGerd Hoffmann g->stats.max_inflight = g->inflight; 78362232bf4SGerd Hoffmann } 784*9d9e1521SGerd Hoffmann fprintf(stderr, "inflight: %3d (+)\r", g->inflight); 785*9d9e1521SGerd Hoffmann } 78662232bf4SGerd Hoffmann cmd = g_new(struct virtio_gpu_ctrl_command, 1); 78762232bf4SGerd Hoffmann } 78862232bf4SGerd Hoffmann } 78962232bf4SGerd Hoffmann g_free(cmd); 790*9d9e1521SGerd Hoffmann 791*9d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 792*9d9e1521SGerd Hoffmann if (g->use_virgl_renderer) { 793*9d9e1521SGerd Hoffmann virtio_gpu_virgl_fence_poll(g); 794*9d9e1521SGerd Hoffmann } 795*9d9e1521SGerd Hoffmann #endif 79662232bf4SGerd Hoffmann } 79762232bf4SGerd Hoffmann 79862232bf4SGerd Hoffmann static void virtio_gpu_ctrl_bh(void *opaque) 79962232bf4SGerd Hoffmann { 80062232bf4SGerd Hoffmann VirtIOGPU *g = opaque; 80162232bf4SGerd Hoffmann virtio_gpu_handle_ctrl(&g->parent_obj, g->ctrl_vq); 80262232bf4SGerd Hoffmann } 80362232bf4SGerd Hoffmann 80462232bf4SGerd Hoffmann static void virtio_gpu_handle_cursor(VirtIODevice *vdev, VirtQueue *vq) 80562232bf4SGerd Hoffmann { 80662232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 80762232bf4SGerd Hoffmann VirtQueueElement elem; 80862232bf4SGerd Hoffmann size_t s; 80962232bf4SGerd Hoffmann struct virtio_gpu_update_cursor cursor_info; 81062232bf4SGerd Hoffmann 81162232bf4SGerd Hoffmann if (!virtio_queue_ready(vq)) { 81262232bf4SGerd Hoffmann return; 81362232bf4SGerd Hoffmann } 81462232bf4SGerd Hoffmann while (virtqueue_pop(vq, &elem)) { 81562232bf4SGerd Hoffmann s = iov_to_buf(elem.out_sg, elem.out_num, 0, 81662232bf4SGerd Hoffmann &cursor_info, sizeof(cursor_info)); 81762232bf4SGerd Hoffmann if (s != sizeof(cursor_info)) { 81862232bf4SGerd Hoffmann qemu_log_mask(LOG_GUEST_ERROR, 81962232bf4SGerd Hoffmann "%s: cursor size incorrect %zu vs %zu\n", 82062232bf4SGerd Hoffmann __func__, s, sizeof(cursor_info)); 82162232bf4SGerd Hoffmann } else { 82262232bf4SGerd Hoffmann update_cursor(g, &cursor_info); 82362232bf4SGerd Hoffmann } 82462232bf4SGerd Hoffmann virtqueue_push(vq, &elem, 0); 82562232bf4SGerd Hoffmann virtio_notify(vdev, vq); 82662232bf4SGerd Hoffmann } 82762232bf4SGerd Hoffmann } 82862232bf4SGerd Hoffmann 82962232bf4SGerd Hoffmann static void virtio_gpu_cursor_bh(void *opaque) 83062232bf4SGerd Hoffmann { 83162232bf4SGerd Hoffmann VirtIOGPU *g = opaque; 83262232bf4SGerd Hoffmann virtio_gpu_handle_cursor(&g->parent_obj, g->cursor_vq); 83362232bf4SGerd Hoffmann } 83462232bf4SGerd Hoffmann 83562232bf4SGerd Hoffmann static void virtio_gpu_invalidate_display(void *opaque) 83662232bf4SGerd Hoffmann { 83762232bf4SGerd Hoffmann } 83862232bf4SGerd Hoffmann 83962232bf4SGerd Hoffmann static void virtio_gpu_update_display(void *opaque) 84062232bf4SGerd Hoffmann { 84162232bf4SGerd Hoffmann } 84262232bf4SGerd Hoffmann 84362232bf4SGerd Hoffmann static void virtio_gpu_text_update(void *opaque, console_ch_t *chardata) 84462232bf4SGerd Hoffmann { 84562232bf4SGerd Hoffmann } 84662232bf4SGerd Hoffmann 84762232bf4SGerd Hoffmann static int virtio_gpu_ui_info(void *opaque, uint32_t idx, QemuUIInfo *info) 84862232bf4SGerd Hoffmann { 84962232bf4SGerd Hoffmann VirtIOGPU *g = opaque; 85062232bf4SGerd Hoffmann 85162232bf4SGerd Hoffmann if (idx > g->conf.max_outputs) { 85262232bf4SGerd Hoffmann return -1; 85362232bf4SGerd Hoffmann } 85462232bf4SGerd Hoffmann 85562232bf4SGerd Hoffmann g->req_state[idx].x = info->xoff; 85662232bf4SGerd Hoffmann g->req_state[idx].y = info->yoff; 85762232bf4SGerd Hoffmann g->req_state[idx].width = info->width; 85862232bf4SGerd Hoffmann g->req_state[idx].height = info->height; 85962232bf4SGerd Hoffmann 86062232bf4SGerd Hoffmann if (info->width && info->height) { 86162232bf4SGerd Hoffmann g->enabled_output_bitmask |= (1 << idx); 86262232bf4SGerd Hoffmann } else { 86362232bf4SGerd Hoffmann g->enabled_output_bitmask &= ~(1 << idx); 86462232bf4SGerd Hoffmann } 86562232bf4SGerd Hoffmann 86662232bf4SGerd Hoffmann /* send event to guest */ 86762232bf4SGerd Hoffmann virtio_gpu_notify_event(g, VIRTIO_GPU_EVENT_DISPLAY); 86862232bf4SGerd Hoffmann return 0; 86962232bf4SGerd Hoffmann } 87062232bf4SGerd Hoffmann 87162232bf4SGerd Hoffmann const GraphicHwOps virtio_gpu_ops = { 87262232bf4SGerd Hoffmann .invalidate = virtio_gpu_invalidate_display, 87362232bf4SGerd Hoffmann .gfx_update = virtio_gpu_update_display, 87462232bf4SGerd Hoffmann .text_update = virtio_gpu_text_update, 87562232bf4SGerd Hoffmann .ui_info = virtio_gpu_ui_info, 87662232bf4SGerd Hoffmann }; 87762232bf4SGerd Hoffmann 87862232bf4SGerd Hoffmann static void virtio_gpu_device_realize(DeviceState *qdev, Error **errp) 87962232bf4SGerd Hoffmann { 88062232bf4SGerd Hoffmann VirtIODevice *vdev = VIRTIO_DEVICE(qdev); 88162232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(qdev); 882*9d9e1521SGerd Hoffmann bool have_virgl; 88362232bf4SGerd Hoffmann int i; 88462232bf4SGerd Hoffmann 88562232bf4SGerd Hoffmann g->config_size = sizeof(struct virtio_gpu_config); 88662232bf4SGerd Hoffmann g->virtio_config.num_scanouts = g->conf.max_outputs; 88762232bf4SGerd Hoffmann virtio_init(VIRTIO_DEVICE(g), "virtio-gpu", VIRTIO_ID_GPU, 88862232bf4SGerd Hoffmann g->config_size); 88962232bf4SGerd Hoffmann 89062232bf4SGerd Hoffmann g->req_state[0].width = 1024; 89162232bf4SGerd Hoffmann g->req_state[0].height = 768; 89262232bf4SGerd Hoffmann 893*9d9e1521SGerd Hoffmann g->use_virgl_renderer = false; 894*9d9e1521SGerd Hoffmann #if !defined(CONFIG_VIRGL) || defined(HOST_WORDS_BIGENDIAN) 895*9d9e1521SGerd Hoffmann have_virgl = false; 896*9d9e1521SGerd Hoffmann #else 897*9d9e1521SGerd Hoffmann have_virgl = display_opengl; 898*9d9e1521SGerd Hoffmann #endif 899*9d9e1521SGerd Hoffmann if (!have_virgl) { 900*9d9e1521SGerd Hoffmann g->conf.flags &= ~(1 << VIRTIO_GPU_FLAG_VIRGL_ENABLED); 901*9d9e1521SGerd Hoffmann } 902*9d9e1521SGerd Hoffmann 903*9d9e1521SGerd Hoffmann if (virtio_gpu_virgl_enabled(g->conf)) { 904*9d9e1521SGerd Hoffmann /* use larger control queue in 3d mode */ 905*9d9e1521SGerd Hoffmann g->ctrl_vq = virtio_add_queue(vdev, 256, virtio_gpu_handle_ctrl_cb); 906*9d9e1521SGerd Hoffmann g->cursor_vq = virtio_add_queue(vdev, 16, virtio_gpu_handle_cursor_cb); 907*9d9e1521SGerd Hoffmann g->virtio_config.num_capsets = 1; 908*9d9e1521SGerd Hoffmann } else { 90962232bf4SGerd Hoffmann g->ctrl_vq = virtio_add_queue(vdev, 64, virtio_gpu_handle_ctrl_cb); 91062232bf4SGerd Hoffmann g->cursor_vq = virtio_add_queue(vdev, 16, virtio_gpu_handle_cursor_cb); 911*9d9e1521SGerd Hoffmann } 91262232bf4SGerd Hoffmann 91362232bf4SGerd Hoffmann g->ctrl_bh = qemu_bh_new(virtio_gpu_ctrl_bh, g); 91462232bf4SGerd Hoffmann g->cursor_bh = qemu_bh_new(virtio_gpu_cursor_bh, g); 91562232bf4SGerd Hoffmann QTAILQ_INIT(&g->reslist); 91662232bf4SGerd Hoffmann QTAILQ_INIT(&g->fenceq); 91762232bf4SGerd Hoffmann 91862232bf4SGerd Hoffmann g->enabled_output_bitmask = 1; 91962232bf4SGerd Hoffmann g->qdev = qdev; 92062232bf4SGerd Hoffmann 92162232bf4SGerd Hoffmann for (i = 0; i < g->conf.max_outputs; i++) { 92262232bf4SGerd Hoffmann g->scanout[i].con = 92362232bf4SGerd Hoffmann graphic_console_init(DEVICE(g), i, &virtio_gpu_ops, g); 92462232bf4SGerd Hoffmann if (i > 0) { 92562232bf4SGerd Hoffmann dpy_gfx_replace_surface(g->scanout[i].con, NULL); 92662232bf4SGerd Hoffmann } 92762232bf4SGerd Hoffmann } 92862232bf4SGerd Hoffmann } 92962232bf4SGerd Hoffmann 93062232bf4SGerd Hoffmann static void virtio_gpu_instance_init(Object *obj) 93162232bf4SGerd Hoffmann { 93262232bf4SGerd Hoffmann } 93362232bf4SGerd Hoffmann 93462232bf4SGerd Hoffmann static void virtio_gpu_reset(VirtIODevice *vdev) 93562232bf4SGerd Hoffmann { 93662232bf4SGerd Hoffmann VirtIOGPU *g = VIRTIO_GPU(vdev); 93762232bf4SGerd Hoffmann struct virtio_gpu_simple_resource *res, *tmp; 93862232bf4SGerd Hoffmann int i; 93962232bf4SGerd Hoffmann 94062232bf4SGerd Hoffmann g->enable = 0; 94162232bf4SGerd Hoffmann 94262232bf4SGerd Hoffmann QTAILQ_FOREACH_SAFE(res, &g->reslist, next, tmp) { 94362232bf4SGerd Hoffmann virtio_gpu_resource_destroy(g, res); 94462232bf4SGerd Hoffmann } 94562232bf4SGerd Hoffmann for (i = 0; i < g->conf.max_outputs; i++) { 94662232bf4SGerd Hoffmann #if 0 94762232bf4SGerd Hoffmann g->req_state[i].x = 0; 94862232bf4SGerd Hoffmann g->req_state[i].y = 0; 94962232bf4SGerd Hoffmann if (i == 0) { 95062232bf4SGerd Hoffmann g->req_state[0].width = 1024; 95162232bf4SGerd Hoffmann g->req_state[0].height = 768; 95262232bf4SGerd Hoffmann } else { 95362232bf4SGerd Hoffmann g->req_state[i].width = 0; 95462232bf4SGerd Hoffmann g->req_state[i].height = 0; 95562232bf4SGerd Hoffmann } 95662232bf4SGerd Hoffmann #endif 95762232bf4SGerd Hoffmann g->scanout[i].resource_id = 0; 95862232bf4SGerd Hoffmann g->scanout[i].width = 0; 95962232bf4SGerd Hoffmann g->scanout[i].height = 0; 96062232bf4SGerd Hoffmann g->scanout[i].x = 0; 96162232bf4SGerd Hoffmann g->scanout[i].y = 0; 96262232bf4SGerd Hoffmann g->scanout[i].ds = NULL; 96362232bf4SGerd Hoffmann } 96462232bf4SGerd Hoffmann g->enabled_output_bitmask = 1; 965*9d9e1521SGerd Hoffmann 966*9d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 967*9d9e1521SGerd Hoffmann if (g->use_virgl_renderer) { 968*9d9e1521SGerd Hoffmann virtio_gpu_virgl_reset(g); 969*9d9e1521SGerd Hoffmann g->use_virgl_renderer = 0; 970*9d9e1521SGerd Hoffmann } 971*9d9e1521SGerd Hoffmann #endif 97262232bf4SGerd Hoffmann } 97362232bf4SGerd Hoffmann 97462232bf4SGerd Hoffmann static Property virtio_gpu_properties[] = { 975b3409a31SGerd Hoffmann DEFINE_PROP_UINT32("max_outputs", VirtIOGPU, conf.max_outputs, 1), 976*9d9e1521SGerd Hoffmann #ifdef CONFIG_VIRGL 977*9d9e1521SGerd Hoffmann DEFINE_PROP_BIT("virgl", VirtIOGPU, conf.flags, 978*9d9e1521SGerd Hoffmann VIRTIO_GPU_FLAG_VIRGL_ENABLED, true), 979*9d9e1521SGerd Hoffmann DEFINE_PROP_BIT("stats", VirtIOGPU, conf.flags, 980*9d9e1521SGerd Hoffmann VIRTIO_GPU_FLAG_STATS_ENABLED, false), 981*9d9e1521SGerd Hoffmann #endif 98262232bf4SGerd Hoffmann DEFINE_PROP_END_OF_LIST(), 98362232bf4SGerd Hoffmann }; 98462232bf4SGerd Hoffmann 98562232bf4SGerd Hoffmann static void virtio_gpu_class_init(ObjectClass *klass, void *data) 98662232bf4SGerd Hoffmann { 98762232bf4SGerd Hoffmann DeviceClass *dc = DEVICE_CLASS(klass); 98862232bf4SGerd Hoffmann VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass); 98962232bf4SGerd Hoffmann 99062232bf4SGerd Hoffmann vdc->realize = virtio_gpu_device_realize; 99162232bf4SGerd Hoffmann vdc->get_config = virtio_gpu_get_config; 99262232bf4SGerd Hoffmann vdc->set_config = virtio_gpu_set_config; 99362232bf4SGerd Hoffmann vdc->get_features = virtio_gpu_get_features; 994*9d9e1521SGerd Hoffmann vdc->set_features = virtio_gpu_set_features; 99562232bf4SGerd Hoffmann 99662232bf4SGerd Hoffmann vdc->reset = virtio_gpu_reset; 99762232bf4SGerd Hoffmann 99862232bf4SGerd Hoffmann dc->props = virtio_gpu_properties; 99962232bf4SGerd Hoffmann } 100062232bf4SGerd Hoffmann 100162232bf4SGerd Hoffmann static const TypeInfo virtio_gpu_info = { 100262232bf4SGerd Hoffmann .name = TYPE_VIRTIO_GPU, 100362232bf4SGerd Hoffmann .parent = TYPE_VIRTIO_DEVICE, 100462232bf4SGerd Hoffmann .instance_size = sizeof(VirtIOGPU), 100562232bf4SGerd Hoffmann .instance_init = virtio_gpu_instance_init, 100662232bf4SGerd Hoffmann .class_init = virtio_gpu_class_init, 100762232bf4SGerd Hoffmann }; 100862232bf4SGerd Hoffmann 100962232bf4SGerd Hoffmann static void virtio_register_types(void) 101062232bf4SGerd Hoffmann { 101162232bf4SGerd Hoffmann type_register_static(&virtio_gpu_info); 101262232bf4SGerd Hoffmann } 101362232bf4SGerd Hoffmann 101462232bf4SGerd Hoffmann type_init(virtio_register_types) 101562232bf4SGerd Hoffmann 101662232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctrl_hdr) != 24); 101762232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_update_cursor) != 56); 101862232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_unref) != 32); 101962232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_create_2d) != 40); 102062232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_set_scanout) != 48); 102162232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_flush) != 48); 102262232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_transfer_to_host_2d) != 56); 102362232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_mem_entry) != 16); 102462232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_attach_backing) != 32); 102562232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_detach_backing) != 32); 102662232bf4SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_display_info) != 408); 1027*9d9e1521SGerd Hoffmann 1028*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_transfer_host_3d) != 72); 1029*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resource_create_3d) != 72); 1030*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_create) != 96); 1031*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_destroy) != 24); 1032*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_ctx_resource) != 32); 1033*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_cmd_submit) != 32); 1034*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_get_capset_info) != 32); 1035*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_capset_info) != 40); 1036*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_get_capset) != 32); 1037*9d9e1521SGerd Hoffmann QEMU_BUILD_BUG_ON(sizeof(struct virtio_gpu_resp_capset) != 24); 1038