| /* |
| * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator |
| * |
| * PAPR Virtual SCSI, aka ibmvscsi |
| * |
| * Copyright (c) 2010,2011 Benjamin Herrenschmidt, IBM Corporation. |
| * |
| * Permission is hereby granted, free of charge, to any person obtaining a copy |
| * of this software and associated documentation files (the "Software"), to deal |
| * in the Software without restriction, including without limitation the rights |
| * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell |
| * copies of the Software, and to permit persons to whom the Software is |
| * furnished to do so, subject to the following conditions: |
| * |
| * The above copyright notice and this permission notice shall be included in |
| * all copies or substantial portions of the Software. |
| * |
| * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
| * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
| * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL |
| * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
| * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, |
| * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN |
| * THE SOFTWARE. |
| * |
| * TODO: |
| * |
| * - Cleanups :-) |
| * - Sort out better how to assign devices to VSCSI instances |
| * - Fix residual counts |
| * - Add indirect descriptors support |
| * - Maybe do autosense (PAPR seems to mandate it, linux doesn't care) |
| */ |
| |
| #include "qemu/osdep.h" |
| #include "qemu/module.h" |
| #include "cpu.h" |
| #include "hw/scsi/scsi.h" |
| #include "migration/vmstate.h" |
| #include "scsi/constants.h" |
| #include "srp.h" |
| #include "hw/ppc/spapr.h" |
| #include "hw/ppc/spapr_vio.h" |
| #include "hw/qdev-properties.h" |
| #include "viosrp.h" |
| #include "trace.h" |
| |
| #include <libfdt.h> |
| |
| /* |
| * Virtual SCSI device |
| */ |
| |
| /* Random numbers */ |
| #define VSCSI_MAX_SECTORS 4096 |
| #define VSCSI_REQ_LIMIT 24 |
| |
| #define SRP_RSP_SENSE_DATA_LEN 18 |
| |
| #define SRP_REPORT_LUNS_WLUN 0xc10100000000000ULL |
| |
| typedef union vscsi_crq { |
| struct viosrp_crq s; |
| uint8_t raw[16]; |
| } vscsi_crq; |
| |
| typedef struct vscsi_req { |
| vscsi_crq crq; |
| union viosrp_iu iu; |
| |
| /* SCSI request tracking */ |
| SCSIRequest *sreq; |
| uint32_t qtag; /* qemu tag != srp tag */ |
| bool active; |
| bool writing; |
| bool dma_error; |
| uint32_t data_len; |
| uint32_t senselen; |
| uint8_t sense[SCSI_SENSE_BUF_SIZE]; |
| |
| /* RDMA related bits */ |
| uint8_t dma_fmt; |
| uint16_t local_desc; |
| uint16_t total_desc; |
| uint16_t cdb_offset; |
| uint16_t cur_desc_num; |
| uint16_t cur_desc_offset; |
| } vscsi_req; |
| |
| #define TYPE_VIO_SPAPR_VSCSI_DEVICE "spapr-vscsi" |
| #define VIO_SPAPR_VSCSI_DEVICE(obj) \ |
| OBJECT_CHECK(VSCSIState, (obj), TYPE_VIO_SPAPR_VSCSI_DEVICE) |
| |
| typedef struct { |
| SpaprVioDevice vdev; |
| SCSIBus bus; |
| vscsi_req reqs[VSCSI_REQ_LIMIT]; |
| } VSCSIState; |
| |
| static struct vscsi_req *vscsi_get_req(VSCSIState *s) |
| { |
| vscsi_req *req; |
| int i; |
| |
| for (i = 0; i < VSCSI_REQ_LIMIT; i++) { |
| req = &s->reqs[i]; |
| if (!req->active) { |
| memset(req, 0, sizeof(*req)); |
| req->qtag = i; |
| req->active = 1; |
| return req; |
| } |
| } |
| return NULL; |
| } |
| |
| static struct vscsi_req *vscsi_find_req(VSCSIState *s, uint64_t srp_tag) |
| { |
| vscsi_req *req; |
| int i; |
| |
| for (i = 0; i < VSCSI_REQ_LIMIT; i++) { |
| req = &s->reqs[i]; |
| if (req->iu.srp.cmd.tag == srp_tag) { |
| return req; |
| } |
| } |
| return NULL; |
| } |
| |
| static void vscsi_put_req(vscsi_req *req) |
| { |
| if (req->sreq != NULL) { |
| scsi_req_unref(req->sreq); |
| } |
| req->sreq = NULL; |
| req->active = 0; |
| } |
| |
| static SCSIDevice *vscsi_device_find(SCSIBus *bus, uint64_t srp_lun, int *lun) |
| { |
| int channel = 0, id = 0; |
| |
| retry: |
| switch (srp_lun >> 62) { |
| case 0: |
| if ((srp_lun >> 56) != 0) { |
| channel = (srp_lun >> 56) & 0x3f; |
| id = (srp_lun >> 48) & 0xff; |
| srp_lun <<= 16; |
| goto retry; |
| } |
| *lun = (srp_lun >> 48) & 0xff; |
| break; |
| |
| case 1: |
| *lun = (srp_lun >> 48) & 0x3fff; |
| break; |
| case 2: |
| channel = (srp_lun >> 53) & 0x7; |
| id = (srp_lun >> 56) & 0x3f; |
| *lun = (srp_lun >> 48) & 0x1f; |
| break; |
| case 3: |
| *lun = -1; |
| return NULL; |
| default: |
| abort(); |
| } |
| |
| return scsi_device_find(bus, channel, id, *lun); |
| } |
| |
| static int vscsi_send_iu(VSCSIState *s, vscsi_req *req, |
| uint64_t length, uint8_t format) |
| { |
| long rc, rc1; |
| |
| /* First copy the SRP */ |
| rc = spapr_vio_dma_write(&s->vdev, req->crq.s.IU_data_ptr, |
| &req->iu, length); |
| if (rc) { |
| fprintf(stderr, "vscsi_send_iu: DMA write failure !\n"); |
| } |
| |
| req->crq.s.valid = 0x80; |
| req->crq.s.format = format; |
| req->crq.s.reserved = 0x00; |
| req->crq.s.timeout = cpu_to_be16(0x0000); |
| req->crq.s.IU_length = cpu_to_be16(length); |
| req->crq.s.IU_data_ptr = req->iu.srp.rsp.tag; /* right byte order */ |
| |
| if (rc == 0) { |
| req->crq.s.status = VIOSRP_OK; |
| } else { |
| req->crq.s.status = VIOSRP_ADAPTER_FAIL; |
| } |
| |
| rc1 = spapr_vio_send_crq(&s->vdev, req->crq.raw); |
| if (rc1) { |
| fprintf(stderr, "vscsi_send_iu: Error sending response\n"); |
| return rc1; |
| } |
| |
| return rc; |
| } |
| |
| static void vscsi_makeup_sense(VSCSIState *s, vscsi_req *req, |
| uint8_t key, uint8_t asc, uint8_t ascq) |
| { |
| req->senselen = SRP_RSP_SENSE_DATA_LEN; |
| |
| /* Valid bit and 'current errors' */ |
| req->sense[0] = (0x1 << 7 | 0x70); |
| /* Sense key */ |
| req->sense[2] = key; |
| /* Additional sense length */ |
| req->sense[7] = 0xa; /* 10 bytes */ |
| /* Additional sense code */ |
| req->sense[12] = asc; |
| req->sense[13] = ascq; |
| } |
| |
| static int vscsi_send_rsp(VSCSIState *s, vscsi_req *req, |
| uint8_t status, int32_t res_in, int32_t res_out) |
| { |
| union viosrp_iu *iu = &req->iu; |
| uint64_t tag = iu->srp.rsp.tag; |
| int total_len = sizeof(iu->srp.rsp); |
| uint8_t sol_not = iu->srp.cmd.sol_not; |
| |
| trace_spapr_vscsi_send_rsp(status, res_in, res_out); |
| |
| memset(iu, 0, sizeof(struct srp_rsp)); |
| iu->srp.rsp.opcode = SRP_RSP; |
| iu->srp.rsp.req_lim_delta = cpu_to_be32(1); |
| iu->srp.rsp.tag = tag; |
| |
| /* Handle residuals */ |
| if (res_in < 0) { |
| iu->srp.rsp.flags |= SRP_RSP_FLAG_DIUNDER; |
| res_in = -res_in; |
| } else if (res_in) { |
| iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER; |
| } |
| if (res_out < 0) { |
| iu->srp.rsp.flags |= SRP_RSP_FLAG_DOUNDER; |
| res_out = -res_out; |
| } else if (res_out) { |
| iu->srp.rsp.flags |= SRP_RSP_FLAG_DOOVER; |
| } |
| iu->srp.rsp.data_in_res_cnt = cpu_to_be32(res_in); |
| iu->srp.rsp.data_out_res_cnt = cpu_to_be32(res_out); |
| |
| /* We don't do response data */ |
| /* iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID; */ |
| iu->srp.rsp.resp_data_len = cpu_to_be32(0); |
| |
| /* Handle success vs. failure */ |
| iu->srp.rsp.status = status; |
| if (status) { |
| iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2; |
| if (req->senselen) { |
| req->iu.srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID; |
| req->iu.srp.rsp.sense_data_len = cpu_to_be32(req->senselen); |
| memcpy(req->iu.srp.rsp.data, req->sense, req->senselen); |
| total_len += req->senselen; |
| } |
| } else { |
| iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1; |
| } |
| |
| vscsi_send_iu(s, req, total_len, VIOSRP_SRP_FORMAT); |
| return 0; |
| } |
| |
| static inline struct srp_direct_buf vscsi_swap_desc(struct srp_direct_buf desc) |
| { |
| desc.va = be64_to_cpu(desc.va); |
| desc.len = be32_to_cpu(desc.len); |
| return desc; |
| } |
| |
| static int vscsi_fetch_desc(VSCSIState *s, struct vscsi_req *req, |
| unsigned n, unsigned buf_offset, |
| struct srp_direct_buf *ret) |
| { |
| struct srp_cmd *cmd = &req->iu.srp.cmd; |
| |
| switch (req->dma_fmt) { |
| case SRP_NO_DATA_DESC: { |
| trace_spapr_vscsi_fetch_desc_no_data(); |
| return 0; |
| } |
| case SRP_DATA_DESC_DIRECT: { |
| memcpy(ret, cmd->add_data + req->cdb_offset, sizeof(*ret)); |
| assert(req->cur_desc_num == 0); |
| trace_spapr_vscsi_fetch_desc_direct(); |
| break; |
| } |
| case SRP_DATA_DESC_INDIRECT: { |
| struct srp_indirect_buf *tmp = (struct srp_indirect_buf *) |
| (cmd->add_data + req->cdb_offset); |
| if (n < req->local_desc) { |
| *ret = tmp->desc_list[n]; |
| trace_spapr_vscsi_fetch_desc_indirect(req->qtag, n, |
| req->local_desc); |
| } else if (n < req->total_desc) { |
| int rc; |
| struct srp_direct_buf tbl_desc = vscsi_swap_desc(tmp->table_desc); |
| unsigned desc_offset = n * sizeof(struct srp_direct_buf); |
| |
| if (desc_offset >= tbl_desc.len) { |
| trace_spapr_vscsi_fetch_desc_out_of_range(n, desc_offset); |
| return -1; |
| } |
| rc = spapr_vio_dma_read(&s->vdev, tbl_desc.va + desc_offset, |
| ret, sizeof(struct srp_direct_buf)); |
| if (rc) { |
| trace_spapr_vscsi_fetch_desc_dma_read_error(rc); |
| return -1; |
| } |
| trace_spapr_vscsi_fetch_desc_indirect_seg_ext(req->qtag, n, |
| req->total_desc, |
| tbl_desc.va, |
| tbl_desc.len); |
| } else { |
| trace_spapr_vscsi_fetch_desc_out_of_desc(); |
| return 0; |
| } |
| break; |
| } |
| default: |
| fprintf(stderr, "VSCSI: Unknown format %x\n", req->dma_fmt); |
| return -1; |
| } |
| |
| *ret = vscsi_swap_desc(*ret); |
| if (buf_offset > ret->len) { |
| trace_spapr_vscsi_fetch_desc_out_of_desc_boundary(buf_offset, |
| req->cur_desc_num, |
| ret->len); |
| return -1; |
| } |
| ret->va += buf_offset; |
| ret->len -= buf_offset; |
| |
| trace_spapr_vscsi_fetch_desc_done(req->cur_desc_num, req->cur_desc_offset, |
| ret->va, ret->len); |
| |
| return ret->len ? 1 : 0; |
| } |
| |
| static int vscsi_srp_direct_data(VSCSIState *s, vscsi_req *req, |
| uint8_t *buf, uint32_t len) |
| { |
| struct srp_direct_buf md; |
| uint32_t llen; |
| int rc = 0; |
| |
| rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md); |
| if (rc < 0) { |
| return -1; |
| } else if (rc == 0) { |
| return 0; |
| } |
| |
| llen = MIN(len, md.len); |
| if (llen) { |
| if (req->writing) { /* writing = to device = reading from memory */ |
| rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen); |
| } else { |
| rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen); |
| } |
| } |
| |
| if (rc) { |
| return -1; |
| } |
| req->cur_desc_offset += llen; |
| |
| return llen; |
| } |
| |
| static int vscsi_srp_indirect_data(VSCSIState *s, vscsi_req *req, |
| uint8_t *buf, uint32_t len) |
| { |
| struct srp_direct_buf md; |
| int rc = 0; |
| uint32_t llen, total = 0; |
| |
| trace_spapr_vscsi_srp_indirect_data(len); |
| |
| /* While we have data ... */ |
| while (len) { |
| rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md); |
| if (rc < 0) { |
| return -1; |
| } else if (rc == 0) { |
| break; |
| } |
| |
| /* Perform transfer */ |
| llen = MIN(len, md.len); |
| if (req->writing) { /* writing = to device = reading from memory */ |
| rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen); |
| } else { |
| rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen); |
| } |
| if (rc) { |
| trace_spapr_vscsi_srp_indirect_data_rw(req->writing, rc); |
| break; |
| } |
| trace_spapr_vscsi_srp_indirect_data_buf(buf[0], buf[1], buf[2], buf[3]); |
| |
| len -= llen; |
| buf += llen; |
| |
| total += llen; |
| |
| /* Update current position in the current descriptor */ |
| req->cur_desc_offset += llen; |
| if (md.len == llen) { |
| /* Go to the next descriptor if the current one finished */ |
| ++req->cur_desc_num; |
| req->cur_desc_offset = 0; |
| } |
| } |
| |
| return rc ? -1 : total; |
| } |
| |
| static int vscsi_srp_transfer_data(VSCSIState *s, vscsi_req *req, |
| int writing, uint8_t *buf, uint32_t len) |
| { |
| int err = 0; |
| |
| switch (req->dma_fmt) { |
| case SRP_NO_DATA_DESC: |
| trace_spapr_vscsi_srp_transfer_data(len); |
| break; |
| case SRP_DATA_DESC_DIRECT: |
| err = vscsi_srp_direct_data(s, req, buf, len); |
| break; |
| case SRP_DATA_DESC_INDIRECT: |
| err = vscsi_srp_indirect_data(s, req, buf, len); |
| break; |
| } |
| return err; |
| } |
| |
| /* Bits from linux srp */ |
| static int data_out_desc_size(struct srp_cmd *cmd) |
| { |
| int size = 0; |
| uint8_t fmt = cmd->buf_fmt >> 4; |
| |
| switch (fmt) { |
| case SRP_NO_DATA_DESC: |
| break; |
| case SRP_DATA_DESC_DIRECT: |
| size = sizeof(struct srp_direct_buf); |
| break; |
| case SRP_DATA_DESC_INDIRECT: |
| size = sizeof(struct srp_indirect_buf) + |
| sizeof(struct srp_direct_buf)*cmd->data_out_desc_cnt; |
| break; |
| default: |
| break; |
| } |
| return size; |
| } |
| |
| static int vscsi_preprocess_desc(vscsi_req *req) |
| { |
| struct srp_cmd *cmd = &req->iu.srp.cmd; |
| |
| req->cdb_offset = cmd->add_cdb_len & ~3; |
| |
| if (req->writing) { |
| req->dma_fmt = cmd->buf_fmt >> 4; |
| } else { |
| req->cdb_offset += data_out_desc_size(cmd); |
| req->dma_fmt = cmd->buf_fmt & ((1U << 4) - 1); |
| } |
| |
| switch (req->dma_fmt) { |
| case SRP_NO_DATA_DESC: |
| break; |
| case SRP_DATA_DESC_DIRECT: |
| req->total_desc = req->local_desc = 1; |
| break; |
| case SRP_DATA_DESC_INDIRECT: { |
| struct srp_indirect_buf *ind_tmp = (struct srp_indirect_buf *) |
| (cmd->add_data + req->cdb_offset); |
| |
| req->total_desc = be32_to_cpu(ind_tmp->table_desc.len) / |
| sizeof(struct srp_direct_buf); |
| req->local_desc = req->writing ? cmd->data_out_desc_cnt : |
| cmd->data_in_desc_cnt; |
| break; |
| } |
| default: |
| fprintf(stderr, |
| "vscsi_preprocess_desc: Unknown format %x\n", req->dma_fmt); |
| return -1; |
| } |
| |
| return 0; |
| } |
| |
| /* Callback to indicate that the SCSI layer has completed a transfer. */ |
| static void vscsi_transfer_data(SCSIRequest *sreq, uint32_t len) |
| { |
| VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); |
| vscsi_req *req = sreq->hba_private; |
| uint8_t *buf; |
| int rc = 0; |
| |
| trace_spapr_vscsi_transfer_data(sreq->tag, len, req); |
| if (req == NULL) { |
| fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag); |
| return; |
| } |
| |
| if (len) { |
| buf = scsi_req_get_buf(sreq); |
| rc = vscsi_srp_transfer_data(s, req, req->writing, buf, len); |
| } |
| if (rc < 0) { |
| fprintf(stderr, "VSCSI: RDMA error rc=%d!\n", rc); |
| req->dma_error = true; |
| scsi_req_cancel(req->sreq); |
| return; |
| } |
| |
| /* Start next chunk */ |
| req->data_len -= rc; |
| scsi_req_continue(sreq); |
| } |
| |
| /* Callback to indicate that the SCSI layer has completed a transfer. */ |
| static void vscsi_command_complete(SCSIRequest *sreq, uint32_t status, size_t resid) |
| { |
| VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); |
| vscsi_req *req = sreq->hba_private; |
| int32_t res_in = 0, res_out = 0; |
| |
| trace_spapr_vscsi_command_complete(sreq->tag, status, req); |
| if (req == NULL) { |
| fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag); |
| return; |
| } |
| |
| if (status == CHECK_CONDITION) { |
| req->senselen = scsi_req_get_sense(req->sreq, req->sense, |
| sizeof(req->sense)); |
| trace_spapr_vscsi_command_complete_sense_data1(req->senselen, |
| req->sense[0], req->sense[1], req->sense[2], req->sense[3], |
| req->sense[4], req->sense[5], req->sense[6], req->sense[7]); |
| trace_spapr_vscsi_command_complete_sense_data2( |
| req->sense[8], req->sense[9], req->sense[10], req->sense[11], |
| req->sense[12], req->sense[13], req->sense[14], req->sense[15]); |
| } |
| |
| trace_spapr_vscsi_command_complete_status(status); |
| if (status == 0) { |
| /* We handle overflows, not underflows for normal commands, |
| * but hopefully nobody cares |
| */ |
| if (req->writing) { |
| res_out = req->data_len; |
| } else { |
| res_in = req->data_len; |
| } |
| } |
| vscsi_send_rsp(s, req, status, res_in, res_out); |
| vscsi_put_req(req); |
| } |
| |
| static void vscsi_request_cancelled(SCSIRequest *sreq) |
| { |
| vscsi_req *req = sreq->hba_private; |
| |
| if (req->dma_error) { |
| VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); |
| |
| vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); |
| vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); |
| } |
| vscsi_put_req(req); |
| } |
| |
| static const VMStateDescription vmstate_spapr_vscsi_req = { |
| .name = "spapr_vscsi_req", |
| .version_id = 1, |
| .minimum_version_id = 1, |
| .fields = (VMStateField[]) { |
| VMSTATE_BUFFER(crq.raw, vscsi_req), |
| VMSTATE_BUFFER(iu.srp.reserved, vscsi_req), |
| VMSTATE_UINT32(qtag, vscsi_req), |
| VMSTATE_BOOL(active, vscsi_req), |
| VMSTATE_UINT32(data_len, vscsi_req), |
| VMSTATE_BOOL(writing, vscsi_req), |
| VMSTATE_UINT32(senselen, vscsi_req), |
| VMSTATE_BUFFER(sense, vscsi_req), |
| VMSTATE_UINT8(dma_fmt, vscsi_req), |
| VMSTATE_UINT16(local_desc, vscsi_req), |
| VMSTATE_UINT16(total_desc, vscsi_req), |
| VMSTATE_UINT16(cdb_offset, vscsi_req), |
| /*Restart SCSI request from the beginning for now */ |
| /*VMSTATE_UINT16(cur_desc_num, vscsi_req), |
| VMSTATE_UINT16(cur_desc_offset, vscsi_req),*/ |
| VMSTATE_END_OF_LIST() |
| }, |
| }; |
| |
| static void vscsi_save_request(QEMUFile *f, SCSIRequest *sreq) |
| { |
| vscsi_req *req = sreq->hba_private; |
| assert(req->active); |
| |
| vmstate_save_state(f, &vmstate_spapr_vscsi_req, req, NULL); |
| |
| trace_spapr_vscsi_save_request(req->qtag, req->cur_desc_num, |
| req->cur_desc_offset); |
| } |
| |
| static void *vscsi_load_request(QEMUFile *f, SCSIRequest *sreq) |
| { |
| SCSIBus *bus = sreq->bus; |
| VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(bus->qbus.parent); |
| vscsi_req *req; |
| int rc; |
| |
| assert(sreq->tag < VSCSI_REQ_LIMIT); |
| req = &s->reqs[sreq->tag]; |
| assert(!req->active); |
| |
| memset(req, 0, sizeof(*req)); |
| rc = vmstate_load_state(f, &vmstate_spapr_vscsi_req, req, 1); |
| if (rc) { |
| fprintf(stderr, "VSCSI: failed loading request tag#%u\n", sreq->tag); |
| return NULL; |
| } |
| assert(req->active); |
| |
| req->sreq = scsi_req_ref(sreq); |
| |
| trace_spapr_vscsi_load_request(req->qtag, req->cur_desc_num, |
| req->cur_desc_offset); |
| |
| return req; |
| } |
| |
| static void vscsi_process_login(VSCSIState *s, vscsi_req *req) |
| { |
| union viosrp_iu *iu = &req->iu; |
| struct srp_login_rsp *rsp = &iu->srp.login_rsp; |
| uint64_t tag = iu->srp.rsp.tag; |
| |
| trace_spapr_vscsi_process_login(); |
| |
| /* TODO handle case that requested size is wrong and |
| * buffer format is wrong |
| */ |
| memset(iu, 0, sizeof(struct srp_login_rsp)); |
| rsp->opcode = SRP_LOGIN_RSP; |
| /* Don't advertise quite as many request as we support to |
| * keep room for management stuff etc... |
| */ |
| rsp->req_lim_delta = cpu_to_be32(VSCSI_REQ_LIMIT-2); |
| rsp->tag = tag; |
| rsp->max_it_iu_len = cpu_to_be32(sizeof(union srp_iu)); |
| rsp->max_ti_iu_len = cpu_to_be32(sizeof(union srp_iu)); |
| /* direct and indirect */ |
| rsp->buf_fmt = cpu_to_be16(SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT); |
| |
| vscsi_send_iu(s, req, sizeof(*rsp), VIOSRP_SRP_FORMAT); |
| } |
| |
| static void vscsi_inquiry_no_target(VSCSIState *s, vscsi_req *req) |
| { |
| uint8_t *cdb = req->iu.srp.cmd.cdb; |
| uint8_t resp_data[36]; |
| int rc, len, alen; |
| |
| /* We don't do EVPD. Also check that page_code is 0 */ |
| if ((cdb[1] & 0x01) || cdb[2] != 0) { |
| /* Send INVALID FIELD IN CDB */ |
| vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0); |
| vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); |
| return; |
| } |
| alen = cdb[3]; |
| alen = (alen << 8) | cdb[4]; |
| len = MIN(alen, 36); |
| |
| /* Fake up inquiry using PQ=3 */ |
| memset(resp_data, 0, 36); |
| resp_data[0] = 0x7f; /* Not capable of supporting a device here */ |
| resp_data[2] = 0x06; /* SPS-4 */ |
| resp_data[3] = 0x02; /* Resp data format */ |
| resp_data[4] = 36 - 5; /* Additional length */ |
| resp_data[7] = 0x10; /* Sync transfers */ |
| memcpy(&resp_data[16], "QEMU EMPTY ", 16); |
| memcpy(&resp_data[8], "QEMU ", 8); |
| |
| req->writing = 0; |
| vscsi_preprocess_desc(req); |
| rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len); |
| if (rc < 0) { |
| vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); |
| vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); |
| } else { |
| vscsi_send_rsp(s, req, 0, 36 - rc, 0); |
| } |
| } |
| |
| static void vscsi_report_luns(VSCSIState *s, vscsi_req *req) |
| { |
| BusChild *kid; |
| int i, len, n, rc; |
| uint8_t *resp_data; |
| bool found_lun0; |
| |
| n = 0; |
| found_lun0 = false; |
| QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) { |
| SCSIDevice *dev = SCSI_DEVICE(kid->child); |
| |
| n += 8; |
| if (dev->channel == 0 && dev->id == 0 && dev->lun == 0) { |
| found_lun0 = true; |
| } |
| } |
| if (!found_lun0) { |
| n += 8; |
| } |
| len = n+8; |
| |
| resp_data = g_malloc0(len); |
| stl_be_p(resp_data, n); |
| i = found_lun0 ? 8 : 16; |
| QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) { |
| DeviceState *qdev = kid->child; |
| SCSIDevice *dev = SCSI_DEVICE(qdev); |
| |
| if (dev->id == 0 && dev->channel == 0) { |
| resp_data[i] = 0; /* Use simple LUN for 0 (SAM5 4.7.7.1) */ |
| } else { |
| resp_data[i] = (2 << 6); /* Otherwise LUN addressing (4.7.7.4) */ |
| } |
| resp_data[i] |= dev->id; |
| resp_data[i+1] = (dev->channel << 5); |
| resp_data[i+1] |= dev->lun; |
| i += 8; |
| } |
| |
| vscsi_preprocess_desc(req); |
| rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len); |
| g_free(resp_data); |
| if (rc < 0) { |
| vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); |
| vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); |
| } else { |
| vscsi_send_rsp(s, req, 0, len - rc, 0); |
| } |
| } |
| |
| static int vscsi_queue_cmd(VSCSIState *s, vscsi_req *req) |
| { |
| union srp_iu *srp = &req->iu.srp; |
| SCSIDevice *sdev; |
| int n, lun; |
| |
| if ((srp->cmd.lun == 0 || be64_to_cpu(srp->cmd.lun) == SRP_REPORT_LUNS_WLUN) |
| && srp->cmd.cdb[0] == REPORT_LUNS) { |
| vscsi_report_luns(s, req); |
| return 0; |
| } |
| |
| sdev = vscsi_device_find(&s->bus, be64_to_cpu(srp->cmd.lun), &lun); |
| if (!sdev) { |
| trace_spapr_vscsi_queue_cmd_no_drive(be64_to_cpu(srp->cmd.lun)); |
| if (srp->cmd.cdb[0] == INQUIRY) { |
| vscsi_inquiry_no_target(s, req); |
| } else { |
| vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0x00); |
| vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); |
| } return 1; |
| } |
| |
| req->sreq = scsi_req_new(sdev, req->qtag, lun, srp->cmd.cdb, req); |
| n = scsi_req_enqueue(req->sreq); |
| |
| trace_spapr_vscsi_queue_cmd(req->qtag, srp->cmd.cdb[0], |
| scsi_command_name(srp->cmd.cdb[0]), lun, n); |
| |
| if (n) { |
| /* Transfer direction must be set before preprocessing the |
| * descriptors |
| */ |
| req->writing = (n < 1); |
| |
| /* Preprocess RDMA descriptors */ |
| vscsi_preprocess_desc(req); |
| |
| /* Get transfer direction and initiate transfer */ |
| if (n > 0) { |
| req->data_len = n; |
| } else if (n < 0) { |
| req->data_len = -n; |
| } |
| scsi_req_continue(req->sreq); |
| } |
| /* Don't touch req here, it may have been recycled already */ |
| |
| return 0; |
| } |
| |
| static int vscsi_process_tsk_mgmt(VSCSIState *s, vscsi_req *req) |
| { |
| union viosrp_iu *iu = &req->iu; |
| vscsi_req *tmpreq; |
| int i, lun = 0, resp = SRP_TSK_MGMT_COMPLETE; |
| SCSIDevice *d; |
| uint64_t tag = iu->srp.rsp.tag; |
| uint8_t sol_not = iu->srp.cmd.sol_not; |
| |
| fprintf(stderr, "vscsi_process_tsk_mgmt %02x\n", |
| iu->srp.tsk_mgmt.tsk_mgmt_func); |
| |
| d = vscsi_device_find(&s->bus, be64_to_cpu(req->iu.srp.tsk_mgmt.lun), &lun); |
| if (!d) { |
| resp = SRP_TSK_MGMT_FIELDS_INVALID; |
| } else { |
| switch (iu->srp.tsk_mgmt.tsk_mgmt_func) { |
| case SRP_TSK_ABORT_TASK: |
| if (d->lun != lun) { |
| resp = SRP_TSK_MGMT_FIELDS_INVALID; |
| break; |
| } |
| |
| tmpreq = vscsi_find_req(s, req->iu.srp.tsk_mgmt.task_tag); |
| if (tmpreq && tmpreq->sreq) { |
| assert(tmpreq->sreq->hba_private); |
| scsi_req_cancel(tmpreq->sreq); |
| } |
| break; |
| |
| case SRP_TSK_LUN_RESET: |
| if (d->lun != lun) { |
| resp = SRP_TSK_MGMT_FIELDS_INVALID; |
| break; |
| } |
| |
| qdev_reset_all(&d->qdev); |
| break; |
| |
| case SRP_TSK_ABORT_TASK_SET: |
| case SRP_TSK_CLEAR_TASK_SET: |
| if (d->lun != lun) { |
| resp = SRP_TSK_MGMT_FIELDS_INVALID; |
| break; |
| } |
| |
| for (i = 0; i < VSCSI_REQ_LIMIT; i++) { |
| tmpreq = &s->reqs[i]; |
| if (tmpreq->iu.srp.cmd.lun != req->iu.srp.tsk_mgmt.lun) { |
| continue; |
| } |
| if (!tmpreq->active || !tmpreq->sreq) { |
| continue; |
| } |
| assert(tmpreq->sreq->hba_private); |
| scsi_req_cancel(tmpreq->sreq); |
| } |
| break; |
| |
| case SRP_TSK_CLEAR_ACA: |
| resp = SRP_TSK_MGMT_NOT_SUPPORTED; |
| break; |
| |
| default: |
| resp = SRP_TSK_MGMT_FIELDS_INVALID; |
| break; |
| } |
| } |
| |
| /* Compose the response here as */ |
| memset(iu, 0, sizeof(struct srp_rsp) + 4); |
| iu->srp.rsp.opcode = SRP_RSP; |
| iu->srp.rsp.req_lim_delta = cpu_to_be32(1); |
| iu->srp.rsp.tag = tag; |
| iu->srp.rsp.flags |= SRP_RSP_FLAG_RSPVALID; |
| iu->srp.rsp.resp_data_len = cpu_to_be32(4); |
| if (resp) { |
| iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2; |
| } else { |
| iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1; |
| } |
| |
| iu->srp.rsp.status = GOOD; |
| iu->srp.rsp.data[3] = resp; |
| |
| vscsi_send_iu(s, req, sizeof(iu->srp.rsp) + 4, VIOSRP_SRP_FORMAT); |
| |
| return 1; |
| } |
| |
| static int vscsi_handle_srp_req(VSCSIState *s, vscsi_req *req) |
| { |
| union srp_iu *srp = &req->iu.srp; |
| int done = 1; |
| uint8_t opcode = srp->rsp.opcode; |
| |
| switch (opcode) { |
| case SRP_LOGIN_REQ: |
| vscsi_process_login(s, req); |
| break; |
| case SRP_TSK_MGMT: |
| done = vscsi_process_tsk_mgmt(s, req); |
| break; |
| case SRP_CMD: |
| done = vscsi_queue_cmd(s, req); |
| break; |
| case SRP_LOGIN_RSP: |
| case SRP_I_LOGOUT: |
| case SRP_T_LOGOUT: |
| case SRP_RSP: |
| case SRP_CRED_REQ: |
| case SRP_CRED_RSP: |
| case SRP_AER_REQ: |
| case SRP_AER_RSP: |
| fprintf(stderr, "VSCSI: Unsupported opcode %02x\n", opcode); |
| break; |
| default: |
| fprintf(stderr, "VSCSI: Unknown type %02x\n", opcode); |
| } |
| |
| return done; |
| } |
| |
| static int vscsi_send_adapter_info(VSCSIState *s, vscsi_req *req) |
| { |
| struct viosrp_adapter_info *sinfo; |
| struct mad_adapter_info_data info; |
| int rc; |
| |
| sinfo = &req->iu.mad.adapter_info; |
| |
| #if 0 /* What for ? */ |
| rc = spapr_vio_dma_read(&s->vdev, be64_to_cpu(sinfo->buffer), |
| &info, be16_to_cpu(sinfo->common.length)); |
| if (rc) { |
| fprintf(stderr, "vscsi_send_adapter_info: DMA read failure !\n"); |
| } |
| #endif |
| memset(&info, 0, sizeof(info)); |
| strcpy(info.srp_version, SRP_VERSION); |
| memcpy(info.partition_name, "qemu", sizeof("qemu")); |
| info.partition_number = cpu_to_be32(0); |
| info.mad_version = cpu_to_be32(1); |
| info.os_type = cpu_to_be32(2); |
| info.port_max_txu[0] = cpu_to_be32(VSCSI_MAX_SECTORS << 9); |
| |
| rc = spapr_vio_dma_write(&s->vdev, be64_to_cpu(sinfo->buffer), |
| &info, be16_to_cpu(sinfo->common.length)); |
| if (rc) { |
| fprintf(stderr, "vscsi_send_adapter_info: DMA write failure !\n"); |
| } |
| |
| sinfo->common.status = rc ? cpu_to_be32(1) : 0; |
| |
| return vscsi_send_iu(s, req, sizeof(*sinfo), VIOSRP_MAD_FORMAT); |
| } |
| |
| static int vscsi_send_capabilities(VSCSIState *s, vscsi_req *req) |
| { |
| struct viosrp_capabilities *vcap; |
| struct capabilities cap = { }; |
| uint16_t len, req_len; |
| uint64_t buffer; |
| int rc; |
| |
| vcap = &req->iu.mad.capabilities; |
| req_len = len = be16_to_cpu(vcap->common.length); |
| buffer = be64_to_cpu(vcap->buffer); |
| if (len > sizeof(cap)) { |
| fprintf(stderr, "vscsi_send_capabilities: capabilities size mismatch !\n"); |
| |
| /* |
| * Just read and populate the structure that is known. |
| * Zero rest of the structure. |
| */ |
| len = sizeof(cap); |
| } |
| rc = spapr_vio_dma_read(&s->vdev, buffer, &cap, len); |
| if (rc) { |
| fprintf(stderr, "vscsi_send_capabilities: DMA read failure !\n"); |
| } |
| |
| /* |
| * Current implementation does not suppport any migration or |
| * reservation capabilities. Construct the response telling the |
| * guest not to use them. |
| */ |
| cap.flags = 0; |
| cap.migration.ecl = 0; |
| cap.reserve.type = 0; |
| cap.migration.common.server_support = 0; |
| cap.reserve.common.server_support = 0; |
| |
| rc = spapr_vio_dma_write(&s->vdev, buffer, &cap, len); |
| if (rc) { |
| fprintf(stderr, "vscsi_send_capabilities: DMA write failure !\n"); |
| } |
| if (req_len > len) { |
| /* |
| * Being paranoid and lets not worry about the error code |
| * here. Actual write of the cap is done above. |
| */ |
| spapr_vio_dma_set(&s->vdev, (buffer + len), 0, (req_len - len)); |
| } |
| vcap->common.status = rc ? cpu_to_be32(1) : 0; |
| return vscsi_send_iu(s, req, sizeof(*vcap), VIOSRP_MAD_FORMAT); |
| } |
| |
| static int vscsi_handle_mad_req(VSCSIState *s, vscsi_req *req) |
| { |
| union mad_iu *mad = &req->iu.mad; |
| bool request_handled = false; |
| uint64_t retlen = 0; |
| |
| switch (be32_to_cpu(mad->empty_iu.common.type)) { |
| case VIOSRP_EMPTY_IU_TYPE: |
| fprintf(stderr, "Unsupported EMPTY MAD IU\n"); |
| retlen = sizeof(mad->empty_iu); |
| break; |
| case VIOSRP_ERROR_LOG_TYPE: |
| fprintf(stderr, "Unsupported ERROR LOG MAD IU\n"); |
| retlen = sizeof(mad->error_log); |
| break; |
| case VIOSRP_ADAPTER_INFO_TYPE: |
| vscsi_send_adapter_info(s, req); |
| request_handled = true; |
| break; |
| case VIOSRP_HOST_CONFIG_TYPE: |
| retlen = sizeof(mad->host_config); |
| break; |
| case VIOSRP_CAPABILITIES_TYPE: |
| vscsi_send_capabilities(s, req); |
| request_handled = true; |
| break; |
| default: |
| fprintf(stderr, "VSCSI: Unknown MAD type %02x\n", |
| be32_to_cpu(mad->empty_iu.common.type)); |
| /* |
| * PAPR+ says that "The length field is set to the length |
| * of the data structure(s) used in the command". |
| * As we did not recognize the request type, put zero there. |
| */ |
| retlen = 0; |
| } |
| |
| if (!request_handled) { |
| mad->empty_iu.common.status = cpu_to_be16(VIOSRP_MAD_NOT_SUPPORTED); |
| vscsi_send_iu(s, req, retlen, VIOSRP_MAD_FORMAT); |
| } |
| |
| return 1; |
| } |
| |
| static void vscsi_got_payload(VSCSIState *s, vscsi_crq *crq) |
| { |
| vscsi_req *req; |
| int done; |
| |
| req = vscsi_get_req(s); |
| if (req == NULL) { |
| fprintf(stderr, "VSCSI: Failed to get a request !\n"); |
| return; |
| } |
| |
| /* We only support a limited number of descriptors, we know |
| * the ibmvscsi driver uses up to 10 max, so it should fit |
| * in our 256 bytes IUs. If not we'll have to increase the size |
| * of the structure. |
| */ |
| if (crq->s.IU_length > sizeof(union viosrp_iu)) { |
| fprintf(stderr, "VSCSI: SRP IU too long (%d bytes) !\n", |
| crq->s.IU_length); |
| vscsi_put_req(req); |
| return; |
| } |
| |
| /* XXX Handle failure differently ? */ |
| if (spapr_vio_dma_read(&s->vdev, crq->s.IU_data_ptr, &req->iu, |
| crq->s.IU_length)) { |
| fprintf(stderr, "vscsi_got_payload: DMA read failure !\n"); |
| vscsi_put_req(req); |
| return; |
| } |
| memcpy(&req->crq, crq, sizeof(vscsi_crq)); |
| |
| if (crq->s.format == VIOSRP_MAD_FORMAT) { |
| done = vscsi_handle_mad_req(s, req); |
| } else { |
| done = vscsi_handle_srp_req(s, req); |
| } |
| |
| if (done) { |
| vscsi_put_req(req); |
| } |
| } |
| |
| |
| static int vscsi_do_crq(struct SpaprVioDevice *dev, uint8_t *crq_data) |
| { |
| VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); |
| vscsi_crq crq; |
| |
| memcpy(crq.raw, crq_data, 16); |
| crq.s.timeout = be16_to_cpu(crq.s.timeout); |
| crq.s.IU_length = be16_to_cpu(crq.s.IU_length); |
| crq.s.IU_data_ptr = be64_to_cpu(crq.s.IU_data_ptr); |
| |
| trace_spapr_vscsi_do_crq(crq.raw[0], crq.raw[1]); |
| |
| switch (crq.s.valid) { |
| case 0xc0: /* Init command/response */ |
| |
| /* Respond to initialization request */ |
| if (crq.s.format == 0x01) { |
| memset(crq.raw, 0, 16); |
| crq.s.valid = 0xc0; |
| crq.s.format = 0x02; |
| spapr_vio_send_crq(dev, crq.raw); |
| } |
| |
| /* Note that in hotplug cases, we might get a 0x02 |
| * as a result of us emitting the init request |
| */ |
| |
| break; |
| case 0xff: /* Link event */ |
| |
| /* Not handled for now */ |
| |
| break; |
| case 0x80: /* Payloads */ |
| switch (crq.s.format) { |
| case VIOSRP_SRP_FORMAT: /* AKA VSCSI request */ |
| case VIOSRP_MAD_FORMAT: /* AKA VSCSI response */ |
| vscsi_got_payload(s, &crq); |
| break; |
| case VIOSRP_OS400_FORMAT: |
| case VIOSRP_AIX_FORMAT: |
| case VIOSRP_LINUX_FORMAT: |
| case VIOSRP_INLINE_FORMAT: |
| fprintf(stderr, "vscsi_do_srq: Unsupported payload format %02x\n", |
| crq.s.format); |
| break; |
| default: |
| fprintf(stderr, "vscsi_do_srq: Unknown payload format %02x\n", |
| crq.s.format); |
| } |
| break; |
| default: |
| fprintf(stderr, "vscsi_do_crq: unknown CRQ %02x %02x ...\n", |
| crq.raw[0], crq.raw[1]); |
| }; |
| |
| return 0; |
| } |
| |
| static const struct SCSIBusInfo vscsi_scsi_info = { |
| .tcq = true, |
| .max_channel = 7, /* logical unit addressing format */ |
| .max_target = 63, |
| .max_lun = 31, |
| |
| .transfer_data = vscsi_transfer_data, |
| .complete = vscsi_command_complete, |
| .cancel = vscsi_request_cancelled, |
| .save_request = vscsi_save_request, |
| .load_request = vscsi_load_request, |
| }; |
| |
| static void spapr_vscsi_reset(SpaprVioDevice *dev) |
| { |
| VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); |
| int i; |
| |
| memset(s->reqs, 0, sizeof(s->reqs)); |
| for (i = 0; i < VSCSI_REQ_LIMIT; i++) { |
| s->reqs[i].qtag = i; |
| } |
| } |
| |
| static void spapr_vscsi_realize(SpaprVioDevice *dev, Error **errp) |
| { |
| VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); |
| |
| dev->crq.SendFunc = vscsi_do_crq; |
| |
| scsi_bus_new(&s->bus, sizeof(s->bus), DEVICE(dev), |
| &vscsi_scsi_info, NULL); |
| } |
| |
| void spapr_vscsi_create(SpaprVioBus *bus) |
| { |
| DeviceState *dev; |
| |
| dev = qdev_create(&bus->bus, "spapr-vscsi"); |
| |
| qdev_init_nofail(dev); |
| scsi_bus_legacy_handle_cmdline(&VIO_SPAPR_VSCSI_DEVICE(dev)->bus); |
| } |
| |
| static int spapr_vscsi_devnode(SpaprVioDevice *dev, void *fdt, int node_off) |
| { |
| int ret; |
| |
| ret = fdt_setprop_cell(fdt, node_off, "#address-cells", 2); |
| if (ret < 0) { |
| return ret; |
| } |
| |
| ret = fdt_setprop_cell(fdt, node_off, "#size-cells", 0); |
| if (ret < 0) { |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| static Property spapr_vscsi_properties[] = { |
| DEFINE_SPAPR_PROPERTIES(VSCSIState, vdev), |
| DEFINE_PROP_END_OF_LIST(), |
| }; |
| |
| static const VMStateDescription vmstate_spapr_vscsi = { |
| .name = "spapr_vscsi", |
| .version_id = 1, |
| .minimum_version_id = 1, |
| .fields = (VMStateField[]) { |
| VMSTATE_SPAPR_VIO(vdev, VSCSIState), |
| /* VSCSI state */ |
| /* ???? */ |
| |
| VMSTATE_END_OF_LIST() |
| }, |
| }; |
| |
| static void spapr_vscsi_class_init(ObjectClass *klass, void *data) |
| { |
| DeviceClass *dc = DEVICE_CLASS(klass); |
| SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass); |
| |
| k->realize = spapr_vscsi_realize; |
| k->reset = spapr_vscsi_reset; |
| k->devnode = spapr_vscsi_devnode; |
| k->dt_name = "v-scsi"; |
| k->dt_type = "vscsi"; |
| k->dt_compatible = "IBM,v-scsi"; |
| k->signal_mask = 0x00000001; |
| set_bit(DEVICE_CATEGORY_STORAGE, dc->categories); |
| dc->props = spapr_vscsi_properties; |
| k->rtce_window_size = 0x10000000; |
| dc->vmsd = &vmstate_spapr_vscsi; |
| } |
| |
| static const TypeInfo spapr_vscsi_info = { |
| .name = TYPE_VIO_SPAPR_VSCSI_DEVICE, |
| .parent = TYPE_VIO_SPAPR_DEVICE, |
| .instance_size = sizeof(VSCSIState), |
| .class_init = spapr_vscsi_class_init, |
| }; |
| |
| static void spapr_vscsi_register_types(void) |
| { |
| type_register_static(&spapr_vscsi_info); |
| } |
| |
| type_init(spapr_vscsi_register_types) |