16e790746SPaolo Bonzini /* 26e790746SPaolo Bonzini * Virtio Block Device 36e790746SPaolo Bonzini * 46e790746SPaolo Bonzini * Copyright IBM, Corp. 2007 56e790746SPaolo Bonzini * 66e790746SPaolo Bonzini * Authors: 76e790746SPaolo Bonzini * Anthony Liguori <aliguori@us.ibm.com> 86e790746SPaolo Bonzini * 96e790746SPaolo Bonzini * This work is licensed under the terms of the GNU GPL, version 2. See 106e790746SPaolo Bonzini * the COPYING file in the top-level directory. 116e790746SPaolo Bonzini * 126e790746SPaolo Bonzini */ 136e790746SPaolo Bonzini 146e790746SPaolo Bonzini #include "qemu-common.h" 15*827805a2SFam Zheng #include "qemu/iov.h" 166e790746SPaolo Bonzini #include "qemu/error-report.h" 176e790746SPaolo Bonzini #include "trace.h" 186e790746SPaolo Bonzini #include "hw/block/block.h" 196e790746SPaolo Bonzini #include "sysemu/blockdev.h" 206e790746SPaolo Bonzini #include "hw/virtio/virtio-blk.h" 216e790746SPaolo Bonzini #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 226e790746SPaolo Bonzini # include "dataplane/virtio-blk.h" 2384db52d0SStefan Hajnoczi # include "migration/migration.h" 246e790746SPaolo Bonzini #endif 256e790746SPaolo Bonzini #include "block/scsi.h" 266e790746SPaolo Bonzini #ifdef __linux__ 276e790746SPaolo Bonzini # include <scsi/sg.h> 286e790746SPaolo Bonzini #endif 296e790746SPaolo Bonzini #include "hw/virtio/virtio-bus.h" 306e790746SPaolo Bonzini 31671ec3f0SFam Zheng static VirtIOBlockReq *virtio_blk_alloc_request(VirtIOBlock *s) 32671ec3f0SFam Zheng { 33671ec3f0SFam Zheng VirtIOBlockReq *req = g_slice_new0(VirtIOBlockReq); 34671ec3f0SFam Zheng req->dev = s; 35671ec3f0SFam Zheng req->elem = g_slice_new0(VirtQueueElement); 36671ec3f0SFam Zheng return req; 37671ec3f0SFam Zheng } 38671ec3f0SFam Zheng 39671ec3f0SFam Zheng static void virtio_blk_free_request(VirtIOBlockReq *req) 40671ec3f0SFam Zheng { 41671ec3f0SFam Zheng if (req) { 42671ec3f0SFam Zheng g_slice_free(VirtQueueElement, req->elem); 43671ec3f0SFam Zheng g_slice_free(VirtIOBlockReq, req); 44671ec3f0SFam Zheng } 45671ec3f0SFam Zheng } 46671ec3f0SFam Zheng 476e790746SPaolo Bonzini static void virtio_blk_req_complete(VirtIOBlockReq *req, int status) 486e790746SPaolo Bonzini { 496e790746SPaolo Bonzini VirtIOBlock *s = req->dev; 506e790746SPaolo Bonzini VirtIODevice *vdev = VIRTIO_DEVICE(s); 516e790746SPaolo Bonzini 526e790746SPaolo Bonzini trace_virtio_blk_req_complete(req, status); 536e790746SPaolo Bonzini 546e790746SPaolo Bonzini stb_p(&req->in->status, status); 55671ec3f0SFam Zheng virtqueue_push(s->vq, req->elem, req->qiov.size + sizeof(*req->in)); 566e790746SPaolo Bonzini virtio_notify(vdev, s->vq); 576e790746SPaolo Bonzini } 586e790746SPaolo Bonzini 596e790746SPaolo Bonzini static int virtio_blk_handle_rw_error(VirtIOBlockReq *req, int error, 606e790746SPaolo Bonzini bool is_read) 616e790746SPaolo Bonzini { 626e790746SPaolo Bonzini BlockErrorAction action = bdrv_get_error_action(req->dev->bs, is_read, error); 636e790746SPaolo Bonzini VirtIOBlock *s = req->dev; 646e790746SPaolo Bonzini 65a589569fSWenchao Xia if (action == BLOCK_ERROR_ACTION_STOP) { 666e790746SPaolo Bonzini req->next = s->rq; 676e790746SPaolo Bonzini s->rq = req; 68a589569fSWenchao Xia } else if (action == BLOCK_ERROR_ACTION_REPORT) { 696e790746SPaolo Bonzini virtio_blk_req_complete(req, VIRTIO_BLK_S_IOERR); 706e790746SPaolo Bonzini bdrv_acct_done(s->bs, &req->acct); 71671ec3f0SFam Zheng virtio_blk_free_request(req); 726e790746SPaolo Bonzini } 736e790746SPaolo Bonzini 746e790746SPaolo Bonzini bdrv_error_action(s->bs, action, is_read, error); 75a589569fSWenchao Xia return action != BLOCK_ERROR_ACTION_IGNORE; 766e790746SPaolo Bonzini } 776e790746SPaolo Bonzini 786e790746SPaolo Bonzini static void virtio_blk_rw_complete(void *opaque, int ret) 796e790746SPaolo Bonzini { 806e790746SPaolo Bonzini VirtIOBlockReq *req = opaque; 816e790746SPaolo Bonzini 826e790746SPaolo Bonzini trace_virtio_blk_rw_complete(req, ret); 836e790746SPaolo Bonzini 846e790746SPaolo Bonzini if (ret) { 85*827805a2SFam Zheng bool is_read = !(ldl_p(&req->out.type) & VIRTIO_BLK_T_OUT); 866e790746SPaolo Bonzini if (virtio_blk_handle_rw_error(req, -ret, is_read)) 876e790746SPaolo Bonzini return; 886e790746SPaolo Bonzini } 896e790746SPaolo Bonzini 906e790746SPaolo Bonzini virtio_blk_req_complete(req, VIRTIO_BLK_S_OK); 916e790746SPaolo Bonzini bdrv_acct_done(req->dev->bs, &req->acct); 92671ec3f0SFam Zheng virtio_blk_free_request(req); 936e790746SPaolo Bonzini } 946e790746SPaolo Bonzini 956e790746SPaolo Bonzini static void virtio_blk_flush_complete(void *opaque, int ret) 966e790746SPaolo Bonzini { 976e790746SPaolo Bonzini VirtIOBlockReq *req = opaque; 986e790746SPaolo Bonzini 996e790746SPaolo Bonzini if (ret) { 1006e790746SPaolo Bonzini if (virtio_blk_handle_rw_error(req, -ret, 0)) { 1016e790746SPaolo Bonzini return; 1026e790746SPaolo Bonzini } 1036e790746SPaolo Bonzini } 1046e790746SPaolo Bonzini 1056e790746SPaolo Bonzini virtio_blk_req_complete(req, VIRTIO_BLK_S_OK); 1066e790746SPaolo Bonzini bdrv_acct_done(req->dev->bs, &req->acct); 107671ec3f0SFam Zheng virtio_blk_free_request(req); 1086e790746SPaolo Bonzini } 1096e790746SPaolo Bonzini 1106e790746SPaolo Bonzini static VirtIOBlockReq *virtio_blk_get_request(VirtIOBlock *s) 1116e790746SPaolo Bonzini { 1126e790746SPaolo Bonzini VirtIOBlockReq *req = virtio_blk_alloc_request(s); 1136e790746SPaolo Bonzini 114671ec3f0SFam Zheng if (!virtqueue_pop(s->vq, req->elem)) { 115671ec3f0SFam Zheng virtio_blk_free_request(req); 1166e790746SPaolo Bonzini return NULL; 1176e790746SPaolo Bonzini } 1186e790746SPaolo Bonzini 1196e790746SPaolo Bonzini return req; 1206e790746SPaolo Bonzini } 1216e790746SPaolo Bonzini 1225a05cbeeSFam Zheng int virtio_blk_handle_scsi_req(VirtIOBlock *blk, 1235a05cbeeSFam Zheng VirtQueueElement *elem) 1246e790746SPaolo Bonzini { 1256e790746SPaolo Bonzini int status = VIRTIO_BLK_S_OK; 1265a05cbeeSFam Zheng struct virtio_scsi_inhdr *scsi = NULL; 1275a05cbeeSFam Zheng #ifdef __linux__ 1285a05cbeeSFam Zheng int i; 1295a05cbeeSFam Zheng struct sg_io_hdr hdr; 1305a05cbeeSFam Zheng #endif 1316e790746SPaolo Bonzini 1326e790746SPaolo Bonzini /* 1336e790746SPaolo Bonzini * We require at least one output segment each for the virtio_blk_outhdr 1346e790746SPaolo Bonzini * and the SCSI command block. 1356e790746SPaolo Bonzini * 1366e790746SPaolo Bonzini * We also at least require the virtio_blk_inhdr, the virtio_scsi_inhdr 1376e790746SPaolo Bonzini * and the sense buffer pointer in the input segments. 1386e790746SPaolo Bonzini */ 1395a05cbeeSFam Zheng if (elem->out_num < 2 || elem->in_num < 3) { 1405a05cbeeSFam Zheng status = VIRTIO_BLK_S_IOERR; 1415a05cbeeSFam Zheng goto fail; 1426e790746SPaolo Bonzini } 1436e790746SPaolo Bonzini 1446e790746SPaolo Bonzini /* 1456e790746SPaolo Bonzini * The scsi inhdr is placed in the second-to-last input segment, just 1466e790746SPaolo Bonzini * before the regular inhdr. 1476e790746SPaolo Bonzini */ 1485a05cbeeSFam Zheng scsi = (void *)elem->in_sg[elem->in_num - 2].iov_base; 1496e790746SPaolo Bonzini 1505a05cbeeSFam Zheng if (!blk->blk.scsi) { 1516e790746SPaolo Bonzini status = VIRTIO_BLK_S_UNSUPP; 1526e790746SPaolo Bonzini goto fail; 1536e790746SPaolo Bonzini } 1546e790746SPaolo Bonzini 1556e790746SPaolo Bonzini /* 1566e790746SPaolo Bonzini * No support for bidirection commands yet. 1576e790746SPaolo Bonzini */ 1585a05cbeeSFam Zheng if (elem->out_num > 2 && elem->in_num > 3) { 1596e790746SPaolo Bonzini status = VIRTIO_BLK_S_UNSUPP; 1606e790746SPaolo Bonzini goto fail; 1616e790746SPaolo Bonzini } 1626e790746SPaolo Bonzini 1636e790746SPaolo Bonzini #ifdef __linux__ 1646e790746SPaolo Bonzini memset(&hdr, 0, sizeof(struct sg_io_hdr)); 1656e790746SPaolo Bonzini hdr.interface_id = 'S'; 1665a05cbeeSFam Zheng hdr.cmd_len = elem->out_sg[1].iov_len; 1675a05cbeeSFam Zheng hdr.cmdp = elem->out_sg[1].iov_base; 1686e790746SPaolo Bonzini hdr.dxfer_len = 0; 1696e790746SPaolo Bonzini 1705a05cbeeSFam Zheng if (elem->out_num > 2) { 1716e790746SPaolo Bonzini /* 1726e790746SPaolo Bonzini * If there are more than the minimally required 2 output segments 1736e790746SPaolo Bonzini * there is write payload starting from the third iovec. 1746e790746SPaolo Bonzini */ 1756e790746SPaolo Bonzini hdr.dxfer_direction = SG_DXFER_TO_DEV; 1765a05cbeeSFam Zheng hdr.iovec_count = elem->out_num - 2; 1776e790746SPaolo Bonzini 1786e790746SPaolo Bonzini for (i = 0; i < hdr.iovec_count; i++) 1795a05cbeeSFam Zheng hdr.dxfer_len += elem->out_sg[i + 2].iov_len; 1806e790746SPaolo Bonzini 1815a05cbeeSFam Zheng hdr.dxferp = elem->out_sg + 2; 1826e790746SPaolo Bonzini 1835a05cbeeSFam Zheng } else if (elem->in_num > 3) { 1846e790746SPaolo Bonzini /* 1856e790746SPaolo Bonzini * If we have more than 3 input segments the guest wants to actually 1866e790746SPaolo Bonzini * read data. 1876e790746SPaolo Bonzini */ 1886e790746SPaolo Bonzini hdr.dxfer_direction = SG_DXFER_FROM_DEV; 1895a05cbeeSFam Zheng hdr.iovec_count = elem->in_num - 3; 1906e790746SPaolo Bonzini for (i = 0; i < hdr.iovec_count; i++) 1915a05cbeeSFam Zheng hdr.dxfer_len += elem->in_sg[i].iov_len; 1926e790746SPaolo Bonzini 1935a05cbeeSFam Zheng hdr.dxferp = elem->in_sg; 1946e790746SPaolo Bonzini } else { 1956e790746SPaolo Bonzini /* 1966e790746SPaolo Bonzini * Some SCSI commands don't actually transfer any data. 1976e790746SPaolo Bonzini */ 1986e790746SPaolo Bonzini hdr.dxfer_direction = SG_DXFER_NONE; 1996e790746SPaolo Bonzini } 2006e790746SPaolo Bonzini 2015a05cbeeSFam Zheng hdr.sbp = elem->in_sg[elem->in_num - 3].iov_base; 2025a05cbeeSFam Zheng hdr.mx_sb_len = elem->in_sg[elem->in_num - 3].iov_len; 2036e790746SPaolo Bonzini 2045a05cbeeSFam Zheng status = bdrv_ioctl(blk->bs, SG_IO, &hdr); 2055a05cbeeSFam Zheng if (status) { 2066e790746SPaolo Bonzini status = VIRTIO_BLK_S_UNSUPP; 2076e790746SPaolo Bonzini goto fail; 2086e790746SPaolo Bonzini } 2096e790746SPaolo Bonzini 2106e790746SPaolo Bonzini /* 2116e790746SPaolo Bonzini * From SCSI-Generic-HOWTO: "Some lower level drivers (e.g. ide-scsi) 2126e790746SPaolo Bonzini * clear the masked_status field [hence status gets cleared too, see 2136e790746SPaolo Bonzini * block/scsi_ioctl.c] even when a CHECK_CONDITION or COMMAND_TERMINATED 2146e790746SPaolo Bonzini * status has occurred. However they do set DRIVER_SENSE in driver_status 2156e790746SPaolo Bonzini * field. Also a (sb_len_wr > 0) indicates there is a sense buffer. 2166e790746SPaolo Bonzini */ 2176e790746SPaolo Bonzini if (hdr.status == 0 && hdr.sb_len_wr > 0) { 2186e790746SPaolo Bonzini hdr.status = CHECK_CONDITION; 2196e790746SPaolo Bonzini } 2206e790746SPaolo Bonzini 2215a05cbeeSFam Zheng stl_p(&scsi->errors, 2226e790746SPaolo Bonzini hdr.status | (hdr.msg_status << 8) | 2236e790746SPaolo Bonzini (hdr.host_status << 16) | (hdr.driver_status << 24)); 2245a05cbeeSFam Zheng stl_p(&scsi->residual, hdr.resid); 2255a05cbeeSFam Zheng stl_p(&scsi->sense_len, hdr.sb_len_wr); 2265a05cbeeSFam Zheng stl_p(&scsi->data_len, hdr.dxfer_len); 2276e790746SPaolo Bonzini 2285a05cbeeSFam Zheng return status; 2296e790746SPaolo Bonzini #else 2306e790746SPaolo Bonzini abort(); 2316e790746SPaolo Bonzini #endif 2326e790746SPaolo Bonzini 2336e790746SPaolo Bonzini fail: 2346e790746SPaolo Bonzini /* Just put anything nonzero so that the ioctl fails in the guest. */ 2355a05cbeeSFam Zheng if (scsi) { 2365a05cbeeSFam Zheng stl_p(&scsi->errors, 255); 2375a05cbeeSFam Zheng } 2385a05cbeeSFam Zheng return status; 2395a05cbeeSFam Zheng } 2405a05cbeeSFam Zheng 2415a05cbeeSFam Zheng static void virtio_blk_handle_scsi(VirtIOBlockReq *req) 2425a05cbeeSFam Zheng { 2435a05cbeeSFam Zheng int status; 2445a05cbeeSFam Zheng 245671ec3f0SFam Zheng status = virtio_blk_handle_scsi_req(req->dev, req->elem); 2466e790746SPaolo Bonzini virtio_blk_req_complete(req, status); 247671ec3f0SFam Zheng virtio_blk_free_request(req); 2486e790746SPaolo Bonzini } 2496e790746SPaolo Bonzini 2506e790746SPaolo Bonzini typedef struct MultiReqBuffer { 2516e790746SPaolo Bonzini BlockRequest blkreq[32]; 2526e790746SPaolo Bonzini unsigned int num_writes; 2536e790746SPaolo Bonzini } MultiReqBuffer; 2546e790746SPaolo Bonzini 2556e790746SPaolo Bonzini static void virtio_submit_multiwrite(BlockDriverState *bs, MultiReqBuffer *mrb) 2566e790746SPaolo Bonzini { 2576e790746SPaolo Bonzini int i, ret; 2586e790746SPaolo Bonzini 2596e790746SPaolo Bonzini if (!mrb->num_writes) { 2606e790746SPaolo Bonzini return; 2616e790746SPaolo Bonzini } 2626e790746SPaolo Bonzini 2636e790746SPaolo Bonzini ret = bdrv_aio_multiwrite(bs, mrb->blkreq, mrb->num_writes); 2646e790746SPaolo Bonzini if (ret != 0) { 2656e790746SPaolo Bonzini for (i = 0; i < mrb->num_writes; i++) { 2666e790746SPaolo Bonzini if (mrb->blkreq[i].error) { 2676e790746SPaolo Bonzini virtio_blk_rw_complete(mrb->blkreq[i].opaque, -EIO); 2686e790746SPaolo Bonzini } 2696e790746SPaolo Bonzini } 2706e790746SPaolo Bonzini } 2716e790746SPaolo Bonzini 2726e790746SPaolo Bonzini mrb->num_writes = 0; 2736e790746SPaolo Bonzini } 2746e790746SPaolo Bonzini 2756e790746SPaolo Bonzini static void virtio_blk_handle_flush(VirtIOBlockReq *req, MultiReqBuffer *mrb) 2766e790746SPaolo Bonzini { 2776e790746SPaolo Bonzini bdrv_acct_start(req->dev->bs, &req->acct, 0, BDRV_ACCT_FLUSH); 2786e790746SPaolo Bonzini 2796e790746SPaolo Bonzini /* 2806e790746SPaolo Bonzini * Make sure all outstanding writes are posted to the backing device. 2816e790746SPaolo Bonzini */ 2826e790746SPaolo Bonzini virtio_submit_multiwrite(req->dev->bs, mrb); 2836e790746SPaolo Bonzini bdrv_aio_flush(req->dev->bs, virtio_blk_flush_complete, req); 2846e790746SPaolo Bonzini } 2856e790746SPaolo Bonzini 2866e790746SPaolo Bonzini static void virtio_blk_handle_write(VirtIOBlockReq *req, MultiReqBuffer *mrb) 2876e790746SPaolo Bonzini { 2886e790746SPaolo Bonzini BlockRequest *blkreq; 2896e790746SPaolo Bonzini uint64_t sector; 2906e790746SPaolo Bonzini 291*827805a2SFam Zheng sector = ldq_p(&req->out.sector); 2926e790746SPaolo Bonzini 2936e790746SPaolo Bonzini bdrv_acct_start(req->dev->bs, &req->acct, req->qiov.size, BDRV_ACCT_WRITE); 2946e790746SPaolo Bonzini 2956e790746SPaolo Bonzini trace_virtio_blk_handle_write(req, sector, req->qiov.size / 512); 2966e790746SPaolo Bonzini 2976e790746SPaolo Bonzini if (sector & req->dev->sector_mask) { 2986e790746SPaolo Bonzini virtio_blk_rw_complete(req, -EIO); 2996e790746SPaolo Bonzini return; 3006e790746SPaolo Bonzini } 3016e790746SPaolo Bonzini if (req->qiov.size % req->dev->conf->logical_block_size) { 3026e790746SPaolo Bonzini virtio_blk_rw_complete(req, -EIO); 3036e790746SPaolo Bonzini return; 3046e790746SPaolo Bonzini } 3056e790746SPaolo Bonzini 3066e790746SPaolo Bonzini if (mrb->num_writes == 32) { 3076e790746SPaolo Bonzini virtio_submit_multiwrite(req->dev->bs, mrb); 3086e790746SPaolo Bonzini } 3096e790746SPaolo Bonzini 3106e790746SPaolo Bonzini blkreq = &mrb->blkreq[mrb->num_writes]; 3116e790746SPaolo Bonzini blkreq->sector = sector; 3126e790746SPaolo Bonzini blkreq->nb_sectors = req->qiov.size / BDRV_SECTOR_SIZE; 3136e790746SPaolo Bonzini blkreq->qiov = &req->qiov; 3146e790746SPaolo Bonzini blkreq->cb = virtio_blk_rw_complete; 3156e790746SPaolo Bonzini blkreq->opaque = req; 3166e790746SPaolo Bonzini blkreq->error = 0; 3176e790746SPaolo Bonzini 3186e790746SPaolo Bonzini mrb->num_writes++; 3196e790746SPaolo Bonzini } 3206e790746SPaolo Bonzini 3216e790746SPaolo Bonzini static void virtio_blk_handle_read(VirtIOBlockReq *req) 3226e790746SPaolo Bonzini { 3236e790746SPaolo Bonzini uint64_t sector; 3246e790746SPaolo Bonzini 325*827805a2SFam Zheng sector = ldq_p(&req->out.sector); 3266e790746SPaolo Bonzini 3276e790746SPaolo Bonzini bdrv_acct_start(req->dev->bs, &req->acct, req->qiov.size, BDRV_ACCT_READ); 3286e790746SPaolo Bonzini 3296e790746SPaolo Bonzini trace_virtio_blk_handle_read(req, sector, req->qiov.size / 512); 3306e790746SPaolo Bonzini 3316e790746SPaolo Bonzini if (sector & req->dev->sector_mask) { 3326e790746SPaolo Bonzini virtio_blk_rw_complete(req, -EIO); 3336e790746SPaolo Bonzini return; 3346e790746SPaolo Bonzini } 3356e790746SPaolo Bonzini if (req->qiov.size % req->dev->conf->logical_block_size) { 3366e790746SPaolo Bonzini virtio_blk_rw_complete(req, -EIO); 3376e790746SPaolo Bonzini return; 3386e790746SPaolo Bonzini } 3396e790746SPaolo Bonzini bdrv_aio_readv(req->dev->bs, sector, &req->qiov, 3406e790746SPaolo Bonzini req->qiov.size / BDRV_SECTOR_SIZE, 3416e790746SPaolo Bonzini virtio_blk_rw_complete, req); 3426e790746SPaolo Bonzini } 3436e790746SPaolo Bonzini 3446e790746SPaolo Bonzini static void virtio_blk_handle_request(VirtIOBlockReq *req, 3456e790746SPaolo Bonzini MultiReqBuffer *mrb) 3466e790746SPaolo Bonzini { 3476e790746SPaolo Bonzini uint32_t type; 348*827805a2SFam Zheng struct iovec *iov = req->elem->out_sg; 349*827805a2SFam Zheng unsigned out_num = req->elem->out_num; 3506e790746SPaolo Bonzini 351671ec3f0SFam Zheng if (req->elem->out_num < 1 || req->elem->in_num < 1) { 3526e790746SPaolo Bonzini error_report("virtio-blk missing headers"); 3536e790746SPaolo Bonzini exit(1); 3546e790746SPaolo Bonzini } 3556e790746SPaolo Bonzini 356*827805a2SFam Zheng if (req->elem->out_sg[0].iov_len < sizeof(req->out) || 357671ec3f0SFam Zheng req->elem->in_sg[req->elem->in_num - 1].iov_len < sizeof(*req->in)) { 3586e790746SPaolo Bonzini error_report("virtio-blk header not in correct element"); 3596e790746SPaolo Bonzini exit(1); 3606e790746SPaolo Bonzini } 3616e790746SPaolo Bonzini 362*827805a2SFam Zheng if (unlikely(iov_to_buf(iov, out_num, 0, &req->out, 363*827805a2SFam Zheng sizeof(req->out)) != sizeof(req->out))) { 364*827805a2SFam Zheng error_report("virtio-blk request outhdr too short"); 365*827805a2SFam Zheng exit(1); 366*827805a2SFam Zheng } 367*827805a2SFam Zheng iov_discard_front(&iov, &out_num, sizeof(req->out)); 368671ec3f0SFam Zheng req->in = (void *)req->elem->in_sg[req->elem->in_num - 1].iov_base; 3696e790746SPaolo Bonzini 370*827805a2SFam Zheng type = ldl_p(&req->out.type); 3716e790746SPaolo Bonzini 3726e790746SPaolo Bonzini if (type & VIRTIO_BLK_T_FLUSH) { 3736e790746SPaolo Bonzini virtio_blk_handle_flush(req, mrb); 3746e790746SPaolo Bonzini } else if (type & VIRTIO_BLK_T_SCSI_CMD) { 3756e790746SPaolo Bonzini virtio_blk_handle_scsi(req); 3766e790746SPaolo Bonzini } else if (type & VIRTIO_BLK_T_GET_ID) { 3776e790746SPaolo Bonzini VirtIOBlock *s = req->dev; 3786e790746SPaolo Bonzini 3796e790746SPaolo Bonzini /* 3806e790746SPaolo Bonzini * NB: per existing s/n string convention the string is 3816e790746SPaolo Bonzini * terminated by '\0' only when shorter than buffer. 3826e790746SPaolo Bonzini */ 383671ec3f0SFam Zheng strncpy(req->elem->in_sg[0].iov_base, 3846e790746SPaolo Bonzini s->blk.serial ? s->blk.serial : "", 385671ec3f0SFam Zheng MIN(req->elem->in_sg[0].iov_len, VIRTIO_BLK_ID_BYTES)); 3866e790746SPaolo Bonzini virtio_blk_req_complete(req, VIRTIO_BLK_S_OK); 387671ec3f0SFam Zheng virtio_blk_free_request(req); 3886e790746SPaolo Bonzini } else if (type & VIRTIO_BLK_T_OUT) { 389671ec3f0SFam Zheng qemu_iovec_init_external(&req->qiov, &req->elem->out_sg[1], 390671ec3f0SFam Zheng req->elem->out_num - 1); 3916e790746SPaolo Bonzini virtio_blk_handle_write(req, mrb); 3926e790746SPaolo Bonzini } else if (type == VIRTIO_BLK_T_IN || type == VIRTIO_BLK_T_BARRIER) { 3936e790746SPaolo Bonzini /* VIRTIO_BLK_T_IN is 0, so we can't just & it. */ 394671ec3f0SFam Zheng qemu_iovec_init_external(&req->qiov, &req->elem->in_sg[0], 395671ec3f0SFam Zheng req->elem->in_num - 1); 3966e790746SPaolo Bonzini virtio_blk_handle_read(req); 3976e790746SPaolo Bonzini } else { 3986e790746SPaolo Bonzini virtio_blk_req_complete(req, VIRTIO_BLK_S_UNSUPP); 399671ec3f0SFam Zheng virtio_blk_free_request(req); 4006e790746SPaolo Bonzini } 4016e790746SPaolo Bonzini } 4026e790746SPaolo Bonzini 4036e790746SPaolo Bonzini static void virtio_blk_handle_output(VirtIODevice *vdev, VirtQueue *vq) 4046e790746SPaolo Bonzini { 4056e790746SPaolo Bonzini VirtIOBlock *s = VIRTIO_BLK(vdev); 4066e790746SPaolo Bonzini VirtIOBlockReq *req; 4076e790746SPaolo Bonzini MultiReqBuffer mrb = { 4086e790746SPaolo Bonzini .num_writes = 0, 4096e790746SPaolo Bonzini }; 4106e790746SPaolo Bonzini 4116e790746SPaolo Bonzini #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 4126e790746SPaolo Bonzini /* Some guests kick before setting VIRTIO_CONFIG_S_DRIVER_OK so start 4136e790746SPaolo Bonzini * dataplane here instead of waiting for .set_status(). 4146e790746SPaolo Bonzini */ 4156e790746SPaolo Bonzini if (s->dataplane) { 4166e790746SPaolo Bonzini virtio_blk_data_plane_start(s->dataplane); 4176e790746SPaolo Bonzini return; 4186e790746SPaolo Bonzini } 4196e790746SPaolo Bonzini #endif 4206e790746SPaolo Bonzini 4216e790746SPaolo Bonzini while ((req = virtio_blk_get_request(s))) { 4226e790746SPaolo Bonzini virtio_blk_handle_request(req, &mrb); 4236e790746SPaolo Bonzini } 4246e790746SPaolo Bonzini 4256e790746SPaolo Bonzini virtio_submit_multiwrite(s->bs, &mrb); 4266e790746SPaolo Bonzini 4276e790746SPaolo Bonzini /* 4286e790746SPaolo Bonzini * FIXME: Want to check for completions before returning to guest mode, 4296e790746SPaolo Bonzini * so cached reads and writes are reported as quickly as possible. But 4306e790746SPaolo Bonzini * that should be done in the generic block layer. 4316e790746SPaolo Bonzini */ 4326e790746SPaolo Bonzini } 4336e790746SPaolo Bonzini 4346e790746SPaolo Bonzini static void virtio_blk_dma_restart_bh(void *opaque) 4356e790746SPaolo Bonzini { 4366e790746SPaolo Bonzini VirtIOBlock *s = opaque; 4376e790746SPaolo Bonzini VirtIOBlockReq *req = s->rq; 4386e790746SPaolo Bonzini MultiReqBuffer mrb = { 4396e790746SPaolo Bonzini .num_writes = 0, 4406e790746SPaolo Bonzini }; 4416e790746SPaolo Bonzini 4426e790746SPaolo Bonzini qemu_bh_delete(s->bh); 4436e790746SPaolo Bonzini s->bh = NULL; 4446e790746SPaolo Bonzini 4456e790746SPaolo Bonzini s->rq = NULL; 4466e790746SPaolo Bonzini 4476e790746SPaolo Bonzini while (req) { 4486e790746SPaolo Bonzini virtio_blk_handle_request(req, &mrb); 4496e790746SPaolo Bonzini req = req->next; 4506e790746SPaolo Bonzini } 4516e790746SPaolo Bonzini 4526e790746SPaolo Bonzini virtio_submit_multiwrite(s->bs, &mrb); 4536e790746SPaolo Bonzini } 4546e790746SPaolo Bonzini 4556e790746SPaolo Bonzini static void virtio_blk_dma_restart_cb(void *opaque, int running, 4566e790746SPaolo Bonzini RunState state) 4576e790746SPaolo Bonzini { 4586e790746SPaolo Bonzini VirtIOBlock *s = opaque; 4596e790746SPaolo Bonzini 4606e790746SPaolo Bonzini if (!running) { 4616e790746SPaolo Bonzini return; 4626e790746SPaolo Bonzini } 4636e790746SPaolo Bonzini 4646e790746SPaolo Bonzini if (!s->bh) { 4656e790746SPaolo Bonzini s->bh = qemu_bh_new(virtio_blk_dma_restart_bh, s); 4666e790746SPaolo Bonzini qemu_bh_schedule(s->bh); 4676e790746SPaolo Bonzini } 4686e790746SPaolo Bonzini } 4696e790746SPaolo Bonzini 4706e790746SPaolo Bonzini static void virtio_blk_reset(VirtIODevice *vdev) 4716e790746SPaolo Bonzini { 4726e790746SPaolo Bonzini VirtIOBlock *s = VIRTIO_BLK(vdev); 4736e790746SPaolo Bonzini 474ef5bc962SPaolo Bonzini #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 4756e790746SPaolo Bonzini if (s->dataplane) { 4766e790746SPaolo Bonzini virtio_blk_data_plane_stop(s->dataplane); 4776e790746SPaolo Bonzini } 4786e790746SPaolo Bonzini #endif 4796e790746SPaolo Bonzini 4806e790746SPaolo Bonzini /* 4816e790746SPaolo Bonzini * This should cancel pending requests, but can't do nicely until there 4826e790746SPaolo Bonzini * are per-device request lists. 4836e790746SPaolo Bonzini */ 4846e790746SPaolo Bonzini bdrv_drain_all(); 485ef5bc962SPaolo Bonzini bdrv_set_enable_write_cache(s->bs, s->original_wce); 4866e790746SPaolo Bonzini } 4876e790746SPaolo Bonzini 4886e790746SPaolo Bonzini /* coalesce internal state, copy to pci i/o region 0 4896e790746SPaolo Bonzini */ 4906e790746SPaolo Bonzini static void virtio_blk_update_config(VirtIODevice *vdev, uint8_t *config) 4916e790746SPaolo Bonzini { 4926e790746SPaolo Bonzini VirtIOBlock *s = VIRTIO_BLK(vdev); 4936e790746SPaolo Bonzini struct virtio_blk_config blkcfg; 4946e790746SPaolo Bonzini uint64_t capacity; 4956e790746SPaolo Bonzini int blk_size = s->conf->logical_block_size; 4966e790746SPaolo Bonzini 4976e790746SPaolo Bonzini bdrv_get_geometry(s->bs, &capacity); 4986e790746SPaolo Bonzini memset(&blkcfg, 0, sizeof(blkcfg)); 4990983979bSPaolo Bonzini stq_p(&blkcfg.capacity, capacity); 5000983979bSPaolo Bonzini stl_p(&blkcfg.seg_max, 128 - 2); 5010983979bSPaolo Bonzini stw_p(&blkcfg.cylinders, s->conf->cyls); 5020983979bSPaolo Bonzini stl_p(&blkcfg.blk_size, blk_size); 5030983979bSPaolo Bonzini stw_p(&blkcfg.min_io_size, s->conf->min_io_size / blk_size); 5040983979bSPaolo Bonzini stw_p(&blkcfg.opt_io_size, s->conf->opt_io_size / blk_size); 5056e790746SPaolo Bonzini blkcfg.heads = s->conf->heads; 5066e790746SPaolo Bonzini /* 5076e790746SPaolo Bonzini * We must ensure that the block device capacity is a multiple of 508e03ba136SPeter Maydell * the logical block size. If that is not the case, let's use 5096e790746SPaolo Bonzini * sector_mask to adopt the geometry to have a correct picture. 5106e790746SPaolo Bonzini * For those devices where the capacity is ok for the given geometry 511e03ba136SPeter Maydell * we don't touch the sector value of the geometry, since some devices 5126e790746SPaolo Bonzini * (like s390 dasd) need a specific value. Here the capacity is already 5136e790746SPaolo Bonzini * cyls*heads*secs*blk_size and the sector value is not block size 5146e790746SPaolo Bonzini * divided by 512 - instead it is the amount of blk_size blocks 5156e790746SPaolo Bonzini * per track (cylinder). 5166e790746SPaolo Bonzini */ 5176e790746SPaolo Bonzini if (bdrv_getlength(s->bs) / s->conf->heads / s->conf->secs % blk_size) { 5186e790746SPaolo Bonzini blkcfg.sectors = s->conf->secs & ~s->sector_mask; 5196e790746SPaolo Bonzini } else { 5206e790746SPaolo Bonzini blkcfg.sectors = s->conf->secs; 5216e790746SPaolo Bonzini } 5226e790746SPaolo Bonzini blkcfg.size_max = 0; 5236e790746SPaolo Bonzini blkcfg.physical_block_exp = get_physical_block_exp(s->conf); 5246e790746SPaolo Bonzini blkcfg.alignment_offset = 0; 5256e790746SPaolo Bonzini blkcfg.wce = bdrv_enable_write_cache(s->bs); 5266e790746SPaolo Bonzini memcpy(config, &blkcfg, sizeof(struct virtio_blk_config)); 5276e790746SPaolo Bonzini } 5286e790746SPaolo Bonzini 5296e790746SPaolo Bonzini static void virtio_blk_set_config(VirtIODevice *vdev, const uint8_t *config) 5306e790746SPaolo Bonzini { 5316e790746SPaolo Bonzini VirtIOBlock *s = VIRTIO_BLK(vdev); 5326e790746SPaolo Bonzini struct virtio_blk_config blkcfg; 5336e790746SPaolo Bonzini 5346e790746SPaolo Bonzini memcpy(&blkcfg, config, sizeof(blkcfg)); 5356d7e73d6SFam Zheng 5366d7e73d6SFam Zheng aio_context_acquire(bdrv_get_aio_context(s->bs)); 5376e790746SPaolo Bonzini bdrv_set_enable_write_cache(s->bs, blkcfg.wce != 0); 5386d7e73d6SFam Zheng aio_context_release(bdrv_get_aio_context(s->bs)); 5396e790746SPaolo Bonzini } 5406e790746SPaolo Bonzini 5416e790746SPaolo Bonzini static uint32_t virtio_blk_get_features(VirtIODevice *vdev, uint32_t features) 5426e790746SPaolo Bonzini { 5436e790746SPaolo Bonzini VirtIOBlock *s = VIRTIO_BLK(vdev); 5446e790746SPaolo Bonzini 5456e790746SPaolo Bonzini features |= (1 << VIRTIO_BLK_F_SEG_MAX); 5466e790746SPaolo Bonzini features |= (1 << VIRTIO_BLK_F_GEOMETRY); 5476e790746SPaolo Bonzini features |= (1 << VIRTIO_BLK_F_TOPOLOGY); 5486e790746SPaolo Bonzini features |= (1 << VIRTIO_BLK_F_BLK_SIZE); 5496e790746SPaolo Bonzini features |= (1 << VIRTIO_BLK_F_SCSI); 5506e790746SPaolo Bonzini 5516e790746SPaolo Bonzini if (s->blk.config_wce) { 5526e790746SPaolo Bonzini features |= (1 << VIRTIO_BLK_F_CONFIG_WCE); 5536e790746SPaolo Bonzini } 5546e790746SPaolo Bonzini if (bdrv_enable_write_cache(s->bs)) 5556e790746SPaolo Bonzini features |= (1 << VIRTIO_BLK_F_WCE); 5566e790746SPaolo Bonzini 5576e790746SPaolo Bonzini if (bdrv_is_read_only(s->bs)) 5586e790746SPaolo Bonzini features |= 1 << VIRTIO_BLK_F_RO; 5596e790746SPaolo Bonzini 5606e790746SPaolo Bonzini return features; 5616e790746SPaolo Bonzini } 5626e790746SPaolo Bonzini 5636e790746SPaolo Bonzini static void virtio_blk_set_status(VirtIODevice *vdev, uint8_t status) 5646e790746SPaolo Bonzini { 5656e790746SPaolo Bonzini VirtIOBlock *s = VIRTIO_BLK(vdev); 5666e790746SPaolo Bonzini uint32_t features; 5676e790746SPaolo Bonzini 5686e790746SPaolo Bonzini #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 5696e790746SPaolo Bonzini if (s->dataplane && !(status & (VIRTIO_CONFIG_S_DRIVER | 5706e790746SPaolo Bonzini VIRTIO_CONFIG_S_DRIVER_OK))) { 5716e790746SPaolo Bonzini virtio_blk_data_plane_stop(s->dataplane); 5726e790746SPaolo Bonzini } 5736e790746SPaolo Bonzini #endif 5746e790746SPaolo Bonzini 5756e790746SPaolo Bonzini if (!(status & VIRTIO_CONFIG_S_DRIVER_OK)) { 5766e790746SPaolo Bonzini return; 5776e790746SPaolo Bonzini } 5786e790746SPaolo Bonzini 5796e790746SPaolo Bonzini features = vdev->guest_features; 580ef5bc962SPaolo Bonzini 581ef5bc962SPaolo Bonzini /* A guest that supports VIRTIO_BLK_F_CONFIG_WCE must be able to send 582ef5bc962SPaolo Bonzini * cache flushes. Thus, the "auto writethrough" behavior is never 583ef5bc962SPaolo Bonzini * necessary for guests that support the VIRTIO_BLK_F_CONFIG_WCE feature. 584ef5bc962SPaolo Bonzini * Leaving it enabled would break the following sequence: 585ef5bc962SPaolo Bonzini * 586ef5bc962SPaolo Bonzini * Guest started with "-drive cache=writethrough" 587ef5bc962SPaolo Bonzini * Guest sets status to 0 588ef5bc962SPaolo Bonzini * Guest sets DRIVER bit in status field 589ef5bc962SPaolo Bonzini * Guest reads host features (WCE=0, CONFIG_WCE=1) 590ef5bc962SPaolo Bonzini * Guest writes guest features (WCE=0, CONFIG_WCE=1) 591ef5bc962SPaolo Bonzini * Guest writes 1 to the WCE configuration field (writeback mode) 592ef5bc962SPaolo Bonzini * Guest sets DRIVER_OK bit in status field 593ef5bc962SPaolo Bonzini * 594ef5bc962SPaolo Bonzini * s->bs would erroneously be placed in writethrough mode. 595ef5bc962SPaolo Bonzini */ 596ef5bc962SPaolo Bonzini if (!(features & (1 << VIRTIO_BLK_F_CONFIG_WCE))) { 5976d7e73d6SFam Zheng aio_context_acquire(bdrv_get_aio_context(s->bs)); 5986d7e73d6SFam Zheng bdrv_set_enable_write_cache(s->bs, 5996d7e73d6SFam Zheng !!(features & (1 << VIRTIO_BLK_F_WCE))); 6006d7e73d6SFam Zheng aio_context_release(bdrv_get_aio_context(s->bs)); 6016e790746SPaolo Bonzini } 602ef5bc962SPaolo Bonzini } 6036e790746SPaolo Bonzini 6046e790746SPaolo Bonzini static void virtio_blk_save(QEMUFile *f, void *opaque) 6056e790746SPaolo Bonzini { 6066e790746SPaolo Bonzini VirtIOBlock *s = opaque; 6076e790746SPaolo Bonzini VirtIODevice *vdev = VIRTIO_DEVICE(s); 6086e790746SPaolo Bonzini VirtIOBlockReq *req = s->rq; 6096e790746SPaolo Bonzini 6106e790746SPaolo Bonzini virtio_save(vdev, f); 6116e790746SPaolo Bonzini 6126e790746SPaolo Bonzini while (req) { 6136e790746SPaolo Bonzini qemu_put_sbyte(f, 1); 614671ec3f0SFam Zheng qemu_put_buffer(f, (unsigned char *)req->elem, 615671ec3f0SFam Zheng sizeof(VirtQueueElement)); 6166e790746SPaolo Bonzini req = req->next; 6176e790746SPaolo Bonzini } 6186e790746SPaolo Bonzini qemu_put_sbyte(f, 0); 6196e790746SPaolo Bonzini } 6206e790746SPaolo Bonzini 6216e790746SPaolo Bonzini static int virtio_blk_load(QEMUFile *f, void *opaque, int version_id) 6226e790746SPaolo Bonzini { 6236e790746SPaolo Bonzini VirtIOBlock *s = opaque; 6246e790746SPaolo Bonzini VirtIODevice *vdev = VIRTIO_DEVICE(s); 6256e790746SPaolo Bonzini int ret; 6266e790746SPaolo Bonzini 6276e790746SPaolo Bonzini if (version_id != 2) 6286e790746SPaolo Bonzini return -EINVAL; 6296e790746SPaolo Bonzini 6306e790746SPaolo Bonzini ret = virtio_load(vdev, f); 6316e790746SPaolo Bonzini if (ret) { 6326e790746SPaolo Bonzini return ret; 6336e790746SPaolo Bonzini } 6346e790746SPaolo Bonzini 6356e790746SPaolo Bonzini while (qemu_get_sbyte(f)) { 6366e790746SPaolo Bonzini VirtIOBlockReq *req = virtio_blk_alloc_request(s); 637671ec3f0SFam Zheng qemu_get_buffer(f, (unsigned char *)req->elem, 638671ec3f0SFam Zheng sizeof(VirtQueueElement)); 6396e790746SPaolo Bonzini req->next = s->rq; 6406e790746SPaolo Bonzini s->rq = req; 6416e790746SPaolo Bonzini 642671ec3f0SFam Zheng virtqueue_map_sg(req->elem->in_sg, req->elem->in_addr, 643671ec3f0SFam Zheng req->elem->in_num, 1); 644671ec3f0SFam Zheng virtqueue_map_sg(req->elem->out_sg, req->elem->out_addr, 645671ec3f0SFam Zheng req->elem->out_num, 0); 6466e790746SPaolo Bonzini } 6476e790746SPaolo Bonzini 6486e790746SPaolo Bonzini return 0; 6496e790746SPaolo Bonzini } 6506e790746SPaolo Bonzini 6516e790746SPaolo Bonzini static void virtio_blk_resize(void *opaque) 6526e790746SPaolo Bonzini { 6536e790746SPaolo Bonzini VirtIODevice *vdev = VIRTIO_DEVICE(opaque); 6546e790746SPaolo Bonzini 6556e790746SPaolo Bonzini virtio_notify_config(vdev); 6566e790746SPaolo Bonzini } 6576e790746SPaolo Bonzini 6586e790746SPaolo Bonzini static const BlockDevOps virtio_block_ops = { 6596e790746SPaolo Bonzini .resize_cb = virtio_blk_resize, 6606e790746SPaolo Bonzini }; 6616e790746SPaolo Bonzini 6626e790746SPaolo Bonzini void virtio_blk_set_conf(DeviceState *dev, VirtIOBlkConf *blk) 6636e790746SPaolo Bonzini { 6646e790746SPaolo Bonzini VirtIOBlock *s = VIRTIO_BLK(dev); 6656e790746SPaolo Bonzini memcpy(&(s->blk), blk, sizeof(struct VirtIOBlkConf)); 6666e790746SPaolo Bonzini } 6676e790746SPaolo Bonzini 66884db52d0SStefan Hajnoczi #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 66984db52d0SStefan Hajnoczi /* Disable dataplane thread during live migration since it does not 67084db52d0SStefan Hajnoczi * update the dirty memory bitmap yet. 67184db52d0SStefan Hajnoczi */ 67284db52d0SStefan Hajnoczi static void virtio_blk_migration_state_changed(Notifier *notifier, void *data) 67384db52d0SStefan Hajnoczi { 67484db52d0SStefan Hajnoczi VirtIOBlock *s = container_of(notifier, VirtIOBlock, 67584db52d0SStefan Hajnoczi migration_state_notifier); 67684db52d0SStefan Hajnoczi MigrationState *mig = data; 6773ffeeef7SAndreas Färber Error *err = NULL; 67884db52d0SStefan Hajnoczi 67984db52d0SStefan Hajnoczi if (migration_in_setup(mig)) { 68084db52d0SStefan Hajnoczi if (!s->dataplane) { 68184db52d0SStefan Hajnoczi return; 68284db52d0SStefan Hajnoczi } 68384db52d0SStefan Hajnoczi virtio_blk_data_plane_destroy(s->dataplane); 68484db52d0SStefan Hajnoczi s->dataplane = NULL; 68584db52d0SStefan Hajnoczi } else if (migration_has_finished(mig) || 68684db52d0SStefan Hajnoczi migration_has_failed(mig)) { 68784db52d0SStefan Hajnoczi if (s->dataplane) { 68884db52d0SStefan Hajnoczi return; 68984db52d0SStefan Hajnoczi } 69084db52d0SStefan Hajnoczi bdrv_drain_all(); /* complete in-flight non-dataplane requests */ 69184db52d0SStefan Hajnoczi virtio_blk_data_plane_create(VIRTIO_DEVICE(s), &s->blk, 6923ffeeef7SAndreas Färber &s->dataplane, &err); 6933ffeeef7SAndreas Färber if (err != NULL) { 6943ffeeef7SAndreas Färber error_report("%s", error_get_pretty(err)); 6953ffeeef7SAndreas Färber error_free(err); 6963ffeeef7SAndreas Färber } 69784db52d0SStefan Hajnoczi } 69884db52d0SStefan Hajnoczi } 69984db52d0SStefan Hajnoczi #endif /* CONFIG_VIRTIO_BLK_DATA_PLANE */ 70084db52d0SStefan Hajnoczi 70175884afdSAndreas Färber static void virtio_blk_device_realize(DeviceState *dev, Error **errp) 7026e790746SPaolo Bonzini { 70375884afdSAndreas Färber VirtIODevice *vdev = VIRTIO_DEVICE(dev); 704179b417eSAndreas Färber VirtIOBlock *s = VIRTIO_BLK(dev); 7056e790746SPaolo Bonzini VirtIOBlkConf *blk = &(s->blk); 7063ffeeef7SAndreas Färber #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 7073ffeeef7SAndreas Färber Error *err = NULL; 7083ffeeef7SAndreas Färber #endif 7096e790746SPaolo Bonzini static int virtio_blk_id; 7106e790746SPaolo Bonzini 7116e790746SPaolo Bonzini if (!blk->conf.bs) { 71275884afdSAndreas Färber error_setg(errp, "drive property not set"); 71375884afdSAndreas Färber return; 7146e790746SPaolo Bonzini } 7156e790746SPaolo Bonzini if (!bdrv_is_inserted(blk->conf.bs)) { 71675884afdSAndreas Färber error_setg(errp, "Device needs media, but drive is empty"); 71775884afdSAndreas Färber return; 7186e790746SPaolo Bonzini } 7196e790746SPaolo Bonzini 7206e790746SPaolo Bonzini blkconf_serial(&blk->conf, &blk->serial); 721ef5bc962SPaolo Bonzini s->original_wce = bdrv_enable_write_cache(blk->conf.bs); 7226e790746SPaolo Bonzini if (blkconf_geometry(&blk->conf, NULL, 65535, 255, 255) < 0) { 72375884afdSAndreas Färber error_setg(errp, "Error setting geometry"); 72475884afdSAndreas Färber return; 7256e790746SPaolo Bonzini } 7266e790746SPaolo Bonzini 7276e790746SPaolo Bonzini virtio_init(vdev, "virtio-blk", VIRTIO_ID_BLOCK, 7286e790746SPaolo Bonzini sizeof(struct virtio_blk_config)); 7296e790746SPaolo Bonzini 7306e790746SPaolo Bonzini s->bs = blk->conf.bs; 7316e790746SPaolo Bonzini s->conf = &blk->conf; 7326e790746SPaolo Bonzini s->rq = NULL; 7336e790746SPaolo Bonzini s->sector_mask = (s->conf->logical_block_size / BDRV_SECTOR_SIZE) - 1; 7346e790746SPaolo Bonzini 7356e790746SPaolo Bonzini s->vq = virtio_add_queue(vdev, 128, virtio_blk_handle_output); 7366e790746SPaolo Bonzini #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 7373ffeeef7SAndreas Färber virtio_blk_data_plane_create(vdev, blk, &s->dataplane, &err); 7383ffeeef7SAndreas Färber if (err != NULL) { 73975884afdSAndreas Färber error_propagate(errp, err); 7406a1a8cc7SKONRAD Frederic virtio_cleanup(vdev); 74175884afdSAndreas Färber return; 7426e790746SPaolo Bonzini } 74384db52d0SStefan Hajnoczi s->migration_state_notifier.notify = virtio_blk_migration_state_changed; 74484db52d0SStefan Hajnoczi add_migration_state_change_notifier(&s->migration_state_notifier); 7456e790746SPaolo Bonzini #endif 7466e790746SPaolo Bonzini 7476e790746SPaolo Bonzini s->change = qemu_add_vm_change_state_handler(virtio_blk_dma_restart_cb, s); 748179b417eSAndreas Färber register_savevm(dev, "virtio-blk", virtio_blk_id++, 2, 7496e790746SPaolo Bonzini virtio_blk_save, virtio_blk_load, s); 7506e790746SPaolo Bonzini bdrv_set_dev_ops(s->bs, &virtio_block_ops, s); 7511b7fd729SPaolo Bonzini bdrv_set_guest_block_size(s->bs, s->conf->logical_block_size); 7526e790746SPaolo Bonzini 7536e790746SPaolo Bonzini bdrv_iostatus_enable(s->bs); 7546e790746SPaolo Bonzini 755179b417eSAndreas Färber add_boot_device_path(s->conf->bootindex, dev, "/disk@0,0"); 7566e790746SPaolo Bonzini } 7576e790746SPaolo Bonzini 758306ec6c3SAndreas Färber static void virtio_blk_device_unrealize(DeviceState *dev, Error **errp) 7596e790746SPaolo Bonzini { 760306ec6c3SAndreas Färber VirtIODevice *vdev = VIRTIO_DEVICE(dev); 761306ec6c3SAndreas Färber VirtIOBlock *s = VIRTIO_BLK(dev); 762306ec6c3SAndreas Färber 7636e790746SPaolo Bonzini #ifdef CONFIG_VIRTIO_BLK_DATA_PLANE 76484db52d0SStefan Hajnoczi remove_migration_state_change_notifier(&s->migration_state_notifier); 7656e790746SPaolo Bonzini virtio_blk_data_plane_destroy(s->dataplane); 7666e790746SPaolo Bonzini s->dataplane = NULL; 7676e790746SPaolo Bonzini #endif 7686e790746SPaolo Bonzini qemu_del_vm_change_state_handler(s->change); 769306ec6c3SAndreas Färber unregister_savevm(dev, "virtio-blk", s); 7706e790746SPaolo Bonzini blockdev_mark_auto_del(s->bs); 7716a1a8cc7SKONRAD Frederic virtio_cleanup(vdev); 7726e790746SPaolo Bonzini } 7736e790746SPaolo Bonzini 7746e790746SPaolo Bonzini static Property virtio_blk_properties[] = { 7756e790746SPaolo Bonzini DEFINE_VIRTIO_BLK_PROPERTIES(VirtIOBlock, blk), 7766e790746SPaolo Bonzini DEFINE_PROP_END_OF_LIST(), 7776e790746SPaolo Bonzini }; 7786e790746SPaolo Bonzini 7796e790746SPaolo Bonzini static void virtio_blk_class_init(ObjectClass *klass, void *data) 7806e790746SPaolo Bonzini { 7816e790746SPaolo Bonzini DeviceClass *dc = DEVICE_CLASS(klass); 7826e790746SPaolo Bonzini VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass); 78375884afdSAndreas Färber 7846e790746SPaolo Bonzini dc->props = virtio_blk_properties; 785125ee0edSMarcel Apfelbaum set_bit(DEVICE_CATEGORY_STORAGE, dc->categories); 78675884afdSAndreas Färber vdc->realize = virtio_blk_device_realize; 787306ec6c3SAndreas Färber vdc->unrealize = virtio_blk_device_unrealize; 7886e790746SPaolo Bonzini vdc->get_config = virtio_blk_update_config; 7896e790746SPaolo Bonzini vdc->set_config = virtio_blk_set_config; 7906e790746SPaolo Bonzini vdc->get_features = virtio_blk_get_features; 7916e790746SPaolo Bonzini vdc->set_status = virtio_blk_set_status; 7926e790746SPaolo Bonzini vdc->reset = virtio_blk_reset; 7936e790746SPaolo Bonzini } 7946e790746SPaolo Bonzini 7956e790746SPaolo Bonzini static const TypeInfo virtio_device_info = { 7966e790746SPaolo Bonzini .name = TYPE_VIRTIO_BLK, 7976e790746SPaolo Bonzini .parent = TYPE_VIRTIO_DEVICE, 7986e790746SPaolo Bonzini .instance_size = sizeof(VirtIOBlock), 7996e790746SPaolo Bonzini .class_init = virtio_blk_class_init, 8006e790746SPaolo Bonzini }; 8016e790746SPaolo Bonzini 8026e790746SPaolo Bonzini static void virtio_register_types(void) 8036e790746SPaolo Bonzini { 8046e790746SPaolo Bonzini type_register_static(&virtio_device_info); 8056e790746SPaolo Bonzini } 8066e790746SPaolo Bonzini 8076e790746SPaolo Bonzini type_init(virtio_register_types) 808