1 /* 2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator 3 * 4 * PAPR Virtual SCSI, aka ibmvscsi 5 * 6 * Copyright (c) 2010,2011 Benjamin Herrenschmidt, IBM Corporation. 7 * 8 * Permission is hereby granted, free of charge, to any person obtaining a copy 9 * of this software and associated documentation files (the "Software"), to deal 10 * in the Software without restriction, including without limitation the rights 11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 12 * copies of the Software, and to permit persons to whom the Software is 13 * furnished to do so, subject to the following conditions: 14 * 15 * The above copyright notice and this permission notice shall be included in 16 * all copies or substantial portions of the Software. 17 * 18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 24 * THE SOFTWARE. 25 * 26 * TODO: 27 * 28 * - Cleanups :-) 29 * - Sort out better how to assign devices to VSCSI instances 30 * - Fix residual counts 31 * - Add indirect descriptors support 32 * - Maybe do autosense (PAPR seems to mandate it, linux doesn't care) 33 */ 34 35 #include "qemu/osdep.h" 36 #include "qemu/module.h" 37 #include "cpu.h" 38 #include "hw/scsi/scsi.h" 39 #include "migration/vmstate.h" 40 #include "scsi/constants.h" 41 #include "srp.h" 42 #include "hw/ppc/spapr.h" 43 #include "hw/ppc/spapr_vio.h" 44 #include "hw/qdev-properties.h" 45 #include "viosrp.h" 46 #include "trace.h" 47 48 #include <libfdt.h> 49 50 /* 51 * Virtual SCSI device 52 */ 53 54 /* Random numbers */ 55 #define VSCSI_MAX_SECTORS 4096 56 #define VSCSI_REQ_LIMIT 24 57 58 #define SRP_RSP_SENSE_DATA_LEN 18 59 60 #define SRP_REPORT_LUNS_WLUN 0xc10100000000000ULL 61 62 typedef union vscsi_crq { 63 struct viosrp_crq s; 64 uint8_t raw[16]; 65 } vscsi_crq; 66 67 typedef struct vscsi_req { 68 vscsi_crq crq; 69 union viosrp_iu iu; 70 71 /* SCSI request tracking */ 72 SCSIRequest *sreq; 73 uint32_t qtag; /* qemu tag != srp tag */ 74 bool active; 75 bool writing; 76 bool dma_error; 77 uint32_t data_len; 78 uint32_t senselen; 79 uint8_t sense[SCSI_SENSE_BUF_SIZE]; 80 81 /* RDMA related bits */ 82 uint8_t dma_fmt; 83 uint16_t local_desc; 84 uint16_t total_desc; 85 uint16_t cdb_offset; 86 uint16_t cur_desc_num; 87 uint16_t cur_desc_offset; 88 } vscsi_req; 89 90 #define TYPE_VIO_SPAPR_VSCSI_DEVICE "spapr-vscsi" 91 #define VIO_SPAPR_VSCSI_DEVICE(obj) \ 92 OBJECT_CHECK(VSCSIState, (obj), TYPE_VIO_SPAPR_VSCSI_DEVICE) 93 94 typedef struct { 95 SpaprVioDevice vdev; 96 SCSIBus bus; 97 vscsi_req reqs[VSCSI_REQ_LIMIT]; 98 } VSCSIState; 99 100 static struct vscsi_req *vscsi_get_req(VSCSIState *s) 101 { 102 vscsi_req *req; 103 int i; 104 105 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 106 req = &s->reqs[i]; 107 if (!req->active) { 108 memset(req, 0, sizeof(*req)); 109 req->qtag = i; 110 req->active = 1; 111 return req; 112 } 113 } 114 return NULL; 115 } 116 117 static struct vscsi_req *vscsi_find_req(VSCSIState *s, uint64_t srp_tag) 118 { 119 vscsi_req *req; 120 int i; 121 122 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 123 req = &s->reqs[i]; 124 if (req->iu.srp.cmd.tag == srp_tag) { 125 return req; 126 } 127 } 128 return NULL; 129 } 130 131 static void vscsi_put_req(vscsi_req *req) 132 { 133 if (req->sreq != NULL) { 134 scsi_req_unref(req->sreq); 135 } 136 req->sreq = NULL; 137 req->active = 0; 138 } 139 140 static SCSIDevice *vscsi_device_find(SCSIBus *bus, uint64_t srp_lun, int *lun) 141 { 142 int channel = 0, id = 0; 143 144 retry: 145 switch (srp_lun >> 62) { 146 case 0: 147 if ((srp_lun >> 56) != 0) { 148 channel = (srp_lun >> 56) & 0x3f; 149 id = (srp_lun >> 48) & 0xff; 150 srp_lun <<= 16; 151 goto retry; 152 } 153 *lun = (srp_lun >> 48) & 0xff; 154 break; 155 156 case 1: 157 *lun = (srp_lun >> 48) & 0x3fff; 158 break; 159 case 2: 160 channel = (srp_lun >> 53) & 0x7; 161 id = (srp_lun >> 56) & 0x3f; 162 *lun = (srp_lun >> 48) & 0x1f; 163 break; 164 case 3: 165 *lun = -1; 166 return NULL; 167 default: 168 abort(); 169 } 170 171 return scsi_device_find(bus, channel, id, *lun); 172 } 173 174 static int vscsi_send_iu(VSCSIState *s, vscsi_req *req, 175 uint64_t length, uint8_t format) 176 { 177 long rc, rc1; 178 179 /* First copy the SRP */ 180 rc = spapr_vio_dma_write(&s->vdev, req->crq.s.IU_data_ptr, 181 &req->iu, length); 182 if (rc) { 183 fprintf(stderr, "vscsi_send_iu: DMA write failure !\n"); 184 } 185 186 req->crq.s.valid = 0x80; 187 req->crq.s.format = format; 188 req->crq.s.reserved = 0x00; 189 req->crq.s.timeout = cpu_to_be16(0x0000); 190 req->crq.s.IU_length = cpu_to_be16(length); 191 req->crq.s.IU_data_ptr = req->iu.srp.rsp.tag; /* right byte order */ 192 193 if (rc == 0) { 194 req->crq.s.status = VIOSRP_OK; 195 } else { 196 req->crq.s.status = VIOSRP_ADAPTER_FAIL; 197 } 198 199 rc1 = spapr_vio_send_crq(&s->vdev, req->crq.raw); 200 if (rc1) { 201 fprintf(stderr, "vscsi_send_iu: Error sending response\n"); 202 return rc1; 203 } 204 205 return rc; 206 } 207 208 static void vscsi_makeup_sense(VSCSIState *s, vscsi_req *req, 209 uint8_t key, uint8_t asc, uint8_t ascq) 210 { 211 req->senselen = SRP_RSP_SENSE_DATA_LEN; 212 213 /* Valid bit and 'current errors' */ 214 req->sense[0] = (0x1 << 7 | 0x70); 215 /* Sense key */ 216 req->sense[2] = key; 217 /* Additional sense length */ 218 req->sense[7] = 0xa; /* 10 bytes */ 219 /* Additional sense code */ 220 req->sense[12] = asc; 221 req->sense[13] = ascq; 222 } 223 224 static int vscsi_send_rsp(VSCSIState *s, vscsi_req *req, 225 uint8_t status, int32_t res_in, int32_t res_out) 226 { 227 union viosrp_iu *iu = &req->iu; 228 uint64_t tag = iu->srp.rsp.tag; 229 int total_len = sizeof(iu->srp.rsp); 230 uint8_t sol_not = iu->srp.cmd.sol_not; 231 232 trace_spapr_vscsi_send_rsp(status, res_in, res_out); 233 234 memset(iu, 0, sizeof(struct srp_rsp)); 235 iu->srp.rsp.opcode = SRP_RSP; 236 iu->srp.rsp.req_lim_delta = cpu_to_be32(1); 237 iu->srp.rsp.tag = tag; 238 239 /* Handle residuals */ 240 if (res_in < 0) { 241 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIUNDER; 242 res_in = -res_in; 243 } else if (res_in) { 244 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER; 245 } 246 if (res_out < 0) { 247 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOUNDER; 248 res_out = -res_out; 249 } else if (res_out) { 250 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOOVER; 251 } 252 iu->srp.rsp.data_in_res_cnt = cpu_to_be32(res_in); 253 iu->srp.rsp.data_out_res_cnt = cpu_to_be32(res_out); 254 255 /* We don't do response data */ 256 /* iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID; */ 257 iu->srp.rsp.resp_data_len = cpu_to_be32(0); 258 259 /* Handle success vs. failure */ 260 iu->srp.rsp.status = status; 261 if (status) { 262 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2; 263 if (req->senselen) { 264 req->iu.srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID; 265 req->iu.srp.rsp.sense_data_len = cpu_to_be32(req->senselen); 266 memcpy(req->iu.srp.rsp.data, req->sense, req->senselen); 267 total_len += req->senselen; 268 } 269 } else { 270 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1; 271 } 272 273 vscsi_send_iu(s, req, total_len, VIOSRP_SRP_FORMAT); 274 return 0; 275 } 276 277 static inline struct srp_direct_buf vscsi_swap_desc(struct srp_direct_buf desc) 278 { 279 desc.va = be64_to_cpu(desc.va); 280 desc.len = be32_to_cpu(desc.len); 281 return desc; 282 } 283 284 static int vscsi_fetch_desc(VSCSIState *s, struct vscsi_req *req, 285 unsigned n, unsigned buf_offset, 286 struct srp_direct_buf *ret) 287 { 288 struct srp_cmd *cmd = &req->iu.srp.cmd; 289 290 switch (req->dma_fmt) { 291 case SRP_NO_DATA_DESC: { 292 trace_spapr_vscsi_fetch_desc_no_data(); 293 return 0; 294 } 295 case SRP_DATA_DESC_DIRECT: { 296 memcpy(ret, cmd->add_data + req->cdb_offset, sizeof(*ret)); 297 assert(req->cur_desc_num == 0); 298 trace_spapr_vscsi_fetch_desc_direct(); 299 break; 300 } 301 case SRP_DATA_DESC_INDIRECT: { 302 struct srp_indirect_buf *tmp = (struct srp_indirect_buf *) 303 (cmd->add_data + req->cdb_offset); 304 if (n < req->local_desc) { 305 *ret = tmp->desc_list[n]; 306 trace_spapr_vscsi_fetch_desc_indirect(req->qtag, n, 307 req->local_desc); 308 } else if (n < req->total_desc) { 309 int rc; 310 struct srp_direct_buf tbl_desc = vscsi_swap_desc(tmp->table_desc); 311 unsigned desc_offset = n * sizeof(struct srp_direct_buf); 312 313 if (desc_offset >= tbl_desc.len) { 314 trace_spapr_vscsi_fetch_desc_out_of_range(n, desc_offset); 315 return -1; 316 } 317 rc = spapr_vio_dma_read(&s->vdev, tbl_desc.va + desc_offset, 318 ret, sizeof(struct srp_direct_buf)); 319 if (rc) { 320 trace_spapr_vscsi_fetch_desc_dma_read_error(rc); 321 return -1; 322 } 323 trace_spapr_vscsi_fetch_desc_indirect_seg_ext(req->qtag, n, 324 req->total_desc, 325 tbl_desc.va, 326 tbl_desc.len); 327 } else { 328 trace_spapr_vscsi_fetch_desc_out_of_desc(); 329 return 0; 330 } 331 break; 332 } 333 default: 334 fprintf(stderr, "VSCSI: Unknown format %x\n", req->dma_fmt); 335 return -1; 336 } 337 338 *ret = vscsi_swap_desc(*ret); 339 if (buf_offset > ret->len) { 340 trace_spapr_vscsi_fetch_desc_out_of_desc_boundary(buf_offset, 341 req->cur_desc_num, 342 ret->len); 343 return -1; 344 } 345 ret->va += buf_offset; 346 ret->len -= buf_offset; 347 348 trace_spapr_vscsi_fetch_desc_done(req->cur_desc_num, req->cur_desc_offset, 349 ret->va, ret->len); 350 351 return ret->len ? 1 : 0; 352 } 353 354 static int vscsi_srp_direct_data(VSCSIState *s, vscsi_req *req, 355 uint8_t *buf, uint32_t len) 356 { 357 struct srp_direct_buf md; 358 uint32_t llen; 359 int rc = 0; 360 361 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md); 362 if (rc < 0) { 363 return -1; 364 } else if (rc == 0) { 365 return 0; 366 } 367 368 llen = MIN(len, md.len); 369 if (llen) { 370 if (req->writing) { /* writing = to device = reading from memory */ 371 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen); 372 } else { 373 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen); 374 } 375 } 376 377 if (rc) { 378 return -1; 379 } 380 req->cur_desc_offset += llen; 381 382 return llen; 383 } 384 385 static int vscsi_srp_indirect_data(VSCSIState *s, vscsi_req *req, 386 uint8_t *buf, uint32_t len) 387 { 388 struct srp_direct_buf md; 389 int rc = 0; 390 uint32_t llen, total = 0; 391 392 trace_spapr_vscsi_srp_indirect_data(len); 393 394 /* While we have data ... */ 395 while (len) { 396 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md); 397 if (rc < 0) { 398 return -1; 399 } else if (rc == 0) { 400 break; 401 } 402 403 /* Perform transfer */ 404 llen = MIN(len, md.len); 405 if (req->writing) { /* writing = to device = reading from memory */ 406 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen); 407 } else { 408 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen); 409 } 410 if (rc) { 411 trace_spapr_vscsi_srp_indirect_data_rw(req->writing, rc); 412 break; 413 } 414 trace_spapr_vscsi_srp_indirect_data_buf(buf[0], buf[1], buf[2], buf[3]); 415 416 len -= llen; 417 buf += llen; 418 419 total += llen; 420 421 /* Update current position in the current descriptor */ 422 req->cur_desc_offset += llen; 423 if (md.len == llen) { 424 /* Go to the next descriptor if the current one finished */ 425 ++req->cur_desc_num; 426 req->cur_desc_offset = 0; 427 } 428 } 429 430 return rc ? -1 : total; 431 } 432 433 static int vscsi_srp_transfer_data(VSCSIState *s, vscsi_req *req, 434 int writing, uint8_t *buf, uint32_t len) 435 { 436 int err = 0; 437 438 switch (req->dma_fmt) { 439 case SRP_NO_DATA_DESC: 440 trace_spapr_vscsi_srp_transfer_data(len); 441 break; 442 case SRP_DATA_DESC_DIRECT: 443 err = vscsi_srp_direct_data(s, req, buf, len); 444 break; 445 case SRP_DATA_DESC_INDIRECT: 446 err = vscsi_srp_indirect_data(s, req, buf, len); 447 break; 448 } 449 return err; 450 } 451 452 /* Bits from linux srp */ 453 static int data_out_desc_size(struct srp_cmd *cmd) 454 { 455 int size = 0; 456 uint8_t fmt = cmd->buf_fmt >> 4; 457 458 switch (fmt) { 459 case SRP_NO_DATA_DESC: 460 break; 461 case SRP_DATA_DESC_DIRECT: 462 size = sizeof(struct srp_direct_buf); 463 break; 464 case SRP_DATA_DESC_INDIRECT: 465 size = sizeof(struct srp_indirect_buf) + 466 sizeof(struct srp_direct_buf)*cmd->data_out_desc_cnt; 467 break; 468 default: 469 break; 470 } 471 return size; 472 } 473 474 static int vscsi_preprocess_desc(vscsi_req *req) 475 { 476 struct srp_cmd *cmd = &req->iu.srp.cmd; 477 478 req->cdb_offset = cmd->add_cdb_len & ~3; 479 480 if (req->writing) { 481 req->dma_fmt = cmd->buf_fmt >> 4; 482 } else { 483 req->cdb_offset += data_out_desc_size(cmd); 484 req->dma_fmt = cmd->buf_fmt & ((1U << 4) - 1); 485 } 486 487 switch (req->dma_fmt) { 488 case SRP_NO_DATA_DESC: 489 break; 490 case SRP_DATA_DESC_DIRECT: 491 req->total_desc = req->local_desc = 1; 492 break; 493 case SRP_DATA_DESC_INDIRECT: { 494 struct srp_indirect_buf *ind_tmp = (struct srp_indirect_buf *) 495 (cmd->add_data + req->cdb_offset); 496 497 req->total_desc = be32_to_cpu(ind_tmp->table_desc.len) / 498 sizeof(struct srp_direct_buf); 499 req->local_desc = req->writing ? cmd->data_out_desc_cnt : 500 cmd->data_in_desc_cnt; 501 break; 502 } 503 default: 504 fprintf(stderr, 505 "vscsi_preprocess_desc: Unknown format %x\n", req->dma_fmt); 506 return -1; 507 } 508 509 return 0; 510 } 511 512 /* Callback to indicate that the SCSI layer has completed a transfer. */ 513 static void vscsi_transfer_data(SCSIRequest *sreq, uint32_t len) 514 { 515 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); 516 vscsi_req *req = sreq->hba_private; 517 uint8_t *buf; 518 int rc = 0; 519 520 trace_spapr_vscsi_transfer_data(sreq->tag, len, req); 521 if (req == NULL) { 522 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag); 523 return; 524 } 525 526 if (len) { 527 buf = scsi_req_get_buf(sreq); 528 rc = vscsi_srp_transfer_data(s, req, req->writing, buf, len); 529 } 530 if (rc < 0) { 531 fprintf(stderr, "VSCSI: RDMA error rc=%d!\n", rc); 532 req->dma_error = true; 533 scsi_req_cancel(req->sreq); 534 return; 535 } 536 537 /* Start next chunk */ 538 req->data_len -= rc; 539 scsi_req_continue(sreq); 540 } 541 542 /* Callback to indicate that the SCSI layer has completed a transfer. */ 543 static void vscsi_command_complete(SCSIRequest *sreq, uint32_t status, size_t resid) 544 { 545 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); 546 vscsi_req *req = sreq->hba_private; 547 int32_t res_in = 0, res_out = 0; 548 549 trace_spapr_vscsi_command_complete(sreq->tag, status, req); 550 if (req == NULL) { 551 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag); 552 return; 553 } 554 555 if (status == CHECK_CONDITION) { 556 req->senselen = scsi_req_get_sense(req->sreq, req->sense, 557 sizeof(req->sense)); 558 trace_spapr_vscsi_command_complete_sense_data1(req->senselen, 559 req->sense[0], req->sense[1], req->sense[2], req->sense[3], 560 req->sense[4], req->sense[5], req->sense[6], req->sense[7]); 561 trace_spapr_vscsi_command_complete_sense_data2( 562 req->sense[8], req->sense[9], req->sense[10], req->sense[11], 563 req->sense[12], req->sense[13], req->sense[14], req->sense[15]); 564 } 565 566 trace_spapr_vscsi_command_complete_status(status); 567 if (status == 0) { 568 /* We handle overflows, not underflows for normal commands, 569 * but hopefully nobody cares 570 */ 571 if (req->writing) { 572 res_out = req->data_len; 573 } else { 574 res_in = req->data_len; 575 } 576 } 577 vscsi_send_rsp(s, req, status, res_in, res_out); 578 vscsi_put_req(req); 579 } 580 581 static void vscsi_request_cancelled(SCSIRequest *sreq) 582 { 583 vscsi_req *req = sreq->hba_private; 584 585 if (req->dma_error) { 586 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); 587 588 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); 589 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 590 } 591 vscsi_put_req(req); 592 } 593 594 static const VMStateDescription vmstate_spapr_vscsi_req = { 595 .name = "spapr_vscsi_req", 596 .version_id = 1, 597 .minimum_version_id = 1, 598 .fields = (VMStateField[]) { 599 VMSTATE_BUFFER(crq.raw, vscsi_req), 600 VMSTATE_BUFFER(iu.srp.reserved, vscsi_req), 601 VMSTATE_UINT32(qtag, vscsi_req), 602 VMSTATE_BOOL(active, vscsi_req), 603 VMSTATE_UINT32(data_len, vscsi_req), 604 VMSTATE_BOOL(writing, vscsi_req), 605 VMSTATE_UINT32(senselen, vscsi_req), 606 VMSTATE_BUFFER(sense, vscsi_req), 607 VMSTATE_UINT8(dma_fmt, vscsi_req), 608 VMSTATE_UINT16(local_desc, vscsi_req), 609 VMSTATE_UINT16(total_desc, vscsi_req), 610 VMSTATE_UINT16(cdb_offset, vscsi_req), 611 /*Restart SCSI request from the beginning for now */ 612 /*VMSTATE_UINT16(cur_desc_num, vscsi_req), 613 VMSTATE_UINT16(cur_desc_offset, vscsi_req),*/ 614 VMSTATE_END_OF_LIST() 615 }, 616 }; 617 618 static void vscsi_save_request(QEMUFile *f, SCSIRequest *sreq) 619 { 620 vscsi_req *req = sreq->hba_private; 621 assert(req->active); 622 623 vmstate_save_state(f, &vmstate_spapr_vscsi_req, req, NULL); 624 625 trace_spapr_vscsi_save_request(req->qtag, req->cur_desc_num, 626 req->cur_desc_offset); 627 } 628 629 static void *vscsi_load_request(QEMUFile *f, SCSIRequest *sreq) 630 { 631 SCSIBus *bus = sreq->bus; 632 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(bus->qbus.parent); 633 vscsi_req *req; 634 int rc; 635 636 assert(sreq->tag < VSCSI_REQ_LIMIT); 637 req = &s->reqs[sreq->tag]; 638 assert(!req->active); 639 640 memset(req, 0, sizeof(*req)); 641 rc = vmstate_load_state(f, &vmstate_spapr_vscsi_req, req, 1); 642 if (rc) { 643 fprintf(stderr, "VSCSI: failed loading request tag#%u\n", sreq->tag); 644 return NULL; 645 } 646 assert(req->active); 647 648 req->sreq = scsi_req_ref(sreq); 649 650 trace_spapr_vscsi_load_request(req->qtag, req->cur_desc_num, 651 req->cur_desc_offset); 652 653 return req; 654 } 655 656 static void vscsi_process_login(VSCSIState *s, vscsi_req *req) 657 { 658 union viosrp_iu *iu = &req->iu; 659 struct srp_login_rsp *rsp = &iu->srp.login_rsp; 660 uint64_t tag = iu->srp.rsp.tag; 661 662 trace_spapr_vscsi_process_login(); 663 664 /* TODO handle case that requested size is wrong and 665 * buffer format is wrong 666 */ 667 memset(iu, 0, sizeof(struct srp_login_rsp)); 668 rsp->opcode = SRP_LOGIN_RSP; 669 /* Don't advertise quite as many request as we support to 670 * keep room for management stuff etc... 671 */ 672 rsp->req_lim_delta = cpu_to_be32(VSCSI_REQ_LIMIT-2); 673 rsp->tag = tag; 674 rsp->max_it_iu_len = cpu_to_be32(sizeof(union srp_iu)); 675 rsp->max_ti_iu_len = cpu_to_be32(sizeof(union srp_iu)); 676 /* direct and indirect */ 677 rsp->buf_fmt = cpu_to_be16(SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT); 678 679 vscsi_send_iu(s, req, sizeof(*rsp), VIOSRP_SRP_FORMAT); 680 } 681 682 static void vscsi_inquiry_no_target(VSCSIState *s, vscsi_req *req) 683 { 684 uint8_t *cdb = req->iu.srp.cmd.cdb; 685 uint8_t resp_data[36]; 686 int rc, len, alen; 687 688 /* We don't do EVPD. Also check that page_code is 0 */ 689 if ((cdb[1] & 0x01) || cdb[2] != 0) { 690 /* Send INVALID FIELD IN CDB */ 691 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0); 692 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 693 return; 694 } 695 alen = cdb[3]; 696 alen = (alen << 8) | cdb[4]; 697 len = MIN(alen, 36); 698 699 /* Fake up inquiry using PQ=3 */ 700 memset(resp_data, 0, 36); 701 resp_data[0] = 0x7f; /* Not capable of supporting a device here */ 702 resp_data[2] = 0x06; /* SPS-4 */ 703 resp_data[3] = 0x02; /* Resp data format */ 704 resp_data[4] = 36 - 5; /* Additional length */ 705 resp_data[7] = 0x10; /* Sync transfers */ 706 memcpy(&resp_data[16], "QEMU EMPTY ", 16); 707 memcpy(&resp_data[8], "QEMU ", 8); 708 709 req->writing = 0; 710 vscsi_preprocess_desc(req); 711 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len); 712 if (rc < 0) { 713 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); 714 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 715 } else { 716 vscsi_send_rsp(s, req, 0, 36 - rc, 0); 717 } 718 } 719 720 static void vscsi_report_luns(VSCSIState *s, vscsi_req *req) 721 { 722 BusChild *kid; 723 int i, len, n, rc; 724 uint8_t *resp_data; 725 bool found_lun0; 726 727 n = 0; 728 found_lun0 = false; 729 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) { 730 SCSIDevice *dev = SCSI_DEVICE(kid->child); 731 732 n += 8; 733 if (dev->channel == 0 && dev->id == 0 && dev->lun == 0) { 734 found_lun0 = true; 735 } 736 } 737 if (!found_lun0) { 738 n += 8; 739 } 740 len = n+8; 741 742 resp_data = g_malloc0(len); 743 stl_be_p(resp_data, n); 744 i = found_lun0 ? 8 : 16; 745 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) { 746 DeviceState *qdev = kid->child; 747 SCSIDevice *dev = SCSI_DEVICE(qdev); 748 749 if (dev->id == 0 && dev->channel == 0) { 750 resp_data[i] = 0; /* Use simple LUN for 0 (SAM5 4.7.7.1) */ 751 } else { 752 resp_data[i] = (2 << 6); /* Otherwise LUN addressing (4.7.7.4) */ 753 } 754 resp_data[i] |= dev->id; 755 resp_data[i+1] = (dev->channel << 5); 756 resp_data[i+1] |= dev->lun; 757 i += 8; 758 } 759 760 vscsi_preprocess_desc(req); 761 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len); 762 g_free(resp_data); 763 if (rc < 0) { 764 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); 765 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 766 } else { 767 vscsi_send_rsp(s, req, 0, len - rc, 0); 768 } 769 } 770 771 static int vscsi_queue_cmd(VSCSIState *s, vscsi_req *req) 772 { 773 union srp_iu *srp = &req->iu.srp; 774 SCSIDevice *sdev; 775 int n, lun; 776 777 if ((srp->cmd.lun == 0 || be64_to_cpu(srp->cmd.lun) == SRP_REPORT_LUNS_WLUN) 778 && srp->cmd.cdb[0] == REPORT_LUNS) { 779 vscsi_report_luns(s, req); 780 return 0; 781 } 782 783 sdev = vscsi_device_find(&s->bus, be64_to_cpu(srp->cmd.lun), &lun); 784 if (!sdev) { 785 trace_spapr_vscsi_queue_cmd_no_drive(be64_to_cpu(srp->cmd.lun)); 786 if (srp->cmd.cdb[0] == INQUIRY) { 787 vscsi_inquiry_no_target(s, req); 788 } else { 789 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0x00); 790 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 791 } return 1; 792 } 793 794 req->sreq = scsi_req_new(sdev, req->qtag, lun, srp->cmd.cdb, req); 795 n = scsi_req_enqueue(req->sreq); 796 797 trace_spapr_vscsi_queue_cmd(req->qtag, srp->cmd.cdb[0], 798 scsi_command_name(srp->cmd.cdb[0]), lun, n); 799 800 if (n) { 801 /* Transfer direction must be set before preprocessing the 802 * descriptors 803 */ 804 req->writing = (n < 1); 805 806 /* Preprocess RDMA descriptors */ 807 vscsi_preprocess_desc(req); 808 809 /* Get transfer direction and initiate transfer */ 810 if (n > 0) { 811 req->data_len = n; 812 } else if (n < 0) { 813 req->data_len = -n; 814 } 815 scsi_req_continue(req->sreq); 816 } 817 /* Don't touch req here, it may have been recycled already */ 818 819 return 0; 820 } 821 822 static int vscsi_process_tsk_mgmt(VSCSIState *s, vscsi_req *req) 823 { 824 union viosrp_iu *iu = &req->iu; 825 vscsi_req *tmpreq; 826 int i, lun = 0, resp = SRP_TSK_MGMT_COMPLETE; 827 SCSIDevice *d; 828 uint64_t tag = iu->srp.rsp.tag; 829 uint8_t sol_not = iu->srp.cmd.sol_not; 830 831 fprintf(stderr, "vscsi_process_tsk_mgmt %02x\n", 832 iu->srp.tsk_mgmt.tsk_mgmt_func); 833 834 d = vscsi_device_find(&s->bus, be64_to_cpu(req->iu.srp.tsk_mgmt.lun), &lun); 835 if (!d) { 836 resp = SRP_TSK_MGMT_FIELDS_INVALID; 837 } else { 838 switch (iu->srp.tsk_mgmt.tsk_mgmt_func) { 839 case SRP_TSK_ABORT_TASK: 840 if (d->lun != lun) { 841 resp = SRP_TSK_MGMT_FIELDS_INVALID; 842 break; 843 } 844 845 tmpreq = vscsi_find_req(s, req->iu.srp.tsk_mgmt.task_tag); 846 if (tmpreq && tmpreq->sreq) { 847 assert(tmpreq->sreq->hba_private); 848 scsi_req_cancel(tmpreq->sreq); 849 } 850 break; 851 852 case SRP_TSK_LUN_RESET: 853 if (d->lun != lun) { 854 resp = SRP_TSK_MGMT_FIELDS_INVALID; 855 break; 856 } 857 858 qdev_reset_all(&d->qdev); 859 break; 860 861 case SRP_TSK_ABORT_TASK_SET: 862 case SRP_TSK_CLEAR_TASK_SET: 863 if (d->lun != lun) { 864 resp = SRP_TSK_MGMT_FIELDS_INVALID; 865 break; 866 } 867 868 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 869 tmpreq = &s->reqs[i]; 870 if (tmpreq->iu.srp.cmd.lun != req->iu.srp.tsk_mgmt.lun) { 871 continue; 872 } 873 if (!tmpreq->active || !tmpreq->sreq) { 874 continue; 875 } 876 assert(tmpreq->sreq->hba_private); 877 scsi_req_cancel(tmpreq->sreq); 878 } 879 break; 880 881 case SRP_TSK_CLEAR_ACA: 882 resp = SRP_TSK_MGMT_NOT_SUPPORTED; 883 break; 884 885 default: 886 resp = SRP_TSK_MGMT_FIELDS_INVALID; 887 break; 888 } 889 } 890 891 /* Compose the response here as */ 892 memset(iu, 0, sizeof(struct srp_rsp) + 4); 893 iu->srp.rsp.opcode = SRP_RSP; 894 iu->srp.rsp.req_lim_delta = cpu_to_be32(1); 895 iu->srp.rsp.tag = tag; 896 iu->srp.rsp.flags |= SRP_RSP_FLAG_RSPVALID; 897 iu->srp.rsp.resp_data_len = cpu_to_be32(4); 898 if (resp) { 899 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2; 900 } else { 901 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1; 902 } 903 904 iu->srp.rsp.status = GOOD; 905 iu->srp.rsp.data[3] = resp; 906 907 vscsi_send_iu(s, req, sizeof(iu->srp.rsp) + 4, VIOSRP_SRP_FORMAT); 908 909 return 1; 910 } 911 912 static int vscsi_handle_srp_req(VSCSIState *s, vscsi_req *req) 913 { 914 union srp_iu *srp = &req->iu.srp; 915 int done = 1; 916 uint8_t opcode = srp->rsp.opcode; 917 918 switch (opcode) { 919 case SRP_LOGIN_REQ: 920 vscsi_process_login(s, req); 921 break; 922 case SRP_TSK_MGMT: 923 done = vscsi_process_tsk_mgmt(s, req); 924 break; 925 case SRP_CMD: 926 done = vscsi_queue_cmd(s, req); 927 break; 928 case SRP_LOGIN_RSP: 929 case SRP_I_LOGOUT: 930 case SRP_T_LOGOUT: 931 case SRP_RSP: 932 case SRP_CRED_REQ: 933 case SRP_CRED_RSP: 934 case SRP_AER_REQ: 935 case SRP_AER_RSP: 936 fprintf(stderr, "VSCSI: Unsupported opcode %02x\n", opcode); 937 break; 938 default: 939 fprintf(stderr, "VSCSI: Unknown type %02x\n", opcode); 940 } 941 942 return done; 943 } 944 945 static int vscsi_send_adapter_info(VSCSIState *s, vscsi_req *req) 946 { 947 struct viosrp_adapter_info *sinfo; 948 struct mad_adapter_info_data info; 949 int rc; 950 951 sinfo = &req->iu.mad.adapter_info; 952 953 #if 0 /* What for ? */ 954 rc = spapr_vio_dma_read(&s->vdev, be64_to_cpu(sinfo->buffer), 955 &info, be16_to_cpu(sinfo->common.length)); 956 if (rc) { 957 fprintf(stderr, "vscsi_send_adapter_info: DMA read failure !\n"); 958 } 959 #endif 960 memset(&info, 0, sizeof(info)); 961 strcpy(info.srp_version, SRP_VERSION); 962 memcpy(info.partition_name, "qemu", sizeof("qemu")); 963 info.partition_number = cpu_to_be32(0); 964 info.mad_version = cpu_to_be32(1); 965 info.os_type = cpu_to_be32(2); 966 info.port_max_txu[0] = cpu_to_be32(VSCSI_MAX_SECTORS << 9); 967 968 rc = spapr_vio_dma_write(&s->vdev, be64_to_cpu(sinfo->buffer), 969 &info, be16_to_cpu(sinfo->common.length)); 970 if (rc) { 971 fprintf(stderr, "vscsi_send_adapter_info: DMA write failure !\n"); 972 } 973 974 sinfo->common.status = rc ? cpu_to_be32(1) : 0; 975 976 return vscsi_send_iu(s, req, sizeof(*sinfo), VIOSRP_MAD_FORMAT); 977 } 978 979 static int vscsi_send_capabilities(VSCSIState *s, vscsi_req *req) 980 { 981 struct viosrp_capabilities *vcap; 982 struct capabilities cap = { }; 983 uint16_t len, req_len; 984 uint64_t buffer; 985 int rc; 986 987 vcap = &req->iu.mad.capabilities; 988 req_len = len = be16_to_cpu(vcap->common.length); 989 buffer = be64_to_cpu(vcap->buffer); 990 if (len > sizeof(cap)) { 991 fprintf(stderr, "vscsi_send_capabilities: capabilities size mismatch !\n"); 992 993 /* 994 * Just read and populate the structure that is known. 995 * Zero rest of the structure. 996 */ 997 len = sizeof(cap); 998 } 999 rc = spapr_vio_dma_read(&s->vdev, buffer, &cap, len); 1000 if (rc) { 1001 fprintf(stderr, "vscsi_send_capabilities: DMA read failure !\n"); 1002 } 1003 1004 /* 1005 * Current implementation does not suppport any migration or 1006 * reservation capabilities. Construct the response telling the 1007 * guest not to use them. 1008 */ 1009 cap.flags = 0; 1010 cap.migration.ecl = 0; 1011 cap.reserve.type = 0; 1012 cap.migration.common.server_support = 0; 1013 cap.reserve.common.server_support = 0; 1014 1015 rc = spapr_vio_dma_write(&s->vdev, buffer, &cap, len); 1016 if (rc) { 1017 fprintf(stderr, "vscsi_send_capabilities: DMA write failure !\n"); 1018 } 1019 if (req_len > len) { 1020 /* 1021 * Being paranoid and lets not worry about the error code 1022 * here. Actual write of the cap is done above. 1023 */ 1024 spapr_vio_dma_set(&s->vdev, (buffer + len), 0, (req_len - len)); 1025 } 1026 vcap->common.status = rc ? cpu_to_be32(1) : 0; 1027 return vscsi_send_iu(s, req, sizeof(*vcap), VIOSRP_MAD_FORMAT); 1028 } 1029 1030 static int vscsi_handle_mad_req(VSCSIState *s, vscsi_req *req) 1031 { 1032 union mad_iu *mad = &req->iu.mad; 1033 bool request_handled = false; 1034 uint64_t retlen = 0; 1035 1036 switch (be32_to_cpu(mad->empty_iu.common.type)) { 1037 case VIOSRP_EMPTY_IU_TYPE: 1038 fprintf(stderr, "Unsupported EMPTY MAD IU\n"); 1039 retlen = sizeof(mad->empty_iu); 1040 break; 1041 case VIOSRP_ERROR_LOG_TYPE: 1042 fprintf(stderr, "Unsupported ERROR LOG MAD IU\n"); 1043 retlen = sizeof(mad->error_log); 1044 break; 1045 case VIOSRP_ADAPTER_INFO_TYPE: 1046 vscsi_send_adapter_info(s, req); 1047 request_handled = true; 1048 break; 1049 case VIOSRP_HOST_CONFIG_TYPE: 1050 retlen = sizeof(mad->host_config); 1051 break; 1052 case VIOSRP_CAPABILITIES_TYPE: 1053 vscsi_send_capabilities(s, req); 1054 request_handled = true; 1055 break; 1056 default: 1057 fprintf(stderr, "VSCSI: Unknown MAD type %02x\n", 1058 be32_to_cpu(mad->empty_iu.common.type)); 1059 /* 1060 * PAPR+ says that "The length field is set to the length 1061 * of the data structure(s) used in the command". 1062 * As we did not recognize the request type, put zero there. 1063 */ 1064 retlen = 0; 1065 } 1066 1067 if (!request_handled) { 1068 mad->empty_iu.common.status = cpu_to_be16(VIOSRP_MAD_NOT_SUPPORTED); 1069 vscsi_send_iu(s, req, retlen, VIOSRP_MAD_FORMAT); 1070 } 1071 1072 return 1; 1073 } 1074 1075 static void vscsi_got_payload(VSCSIState *s, vscsi_crq *crq) 1076 { 1077 vscsi_req *req; 1078 int done; 1079 1080 req = vscsi_get_req(s); 1081 if (req == NULL) { 1082 fprintf(stderr, "VSCSI: Failed to get a request !\n"); 1083 return; 1084 } 1085 1086 /* We only support a limited number of descriptors, we know 1087 * the ibmvscsi driver uses up to 10 max, so it should fit 1088 * in our 256 bytes IUs. If not we'll have to increase the size 1089 * of the structure. 1090 */ 1091 if (crq->s.IU_length > sizeof(union viosrp_iu)) { 1092 fprintf(stderr, "VSCSI: SRP IU too long (%d bytes) !\n", 1093 crq->s.IU_length); 1094 vscsi_put_req(req); 1095 return; 1096 } 1097 1098 /* XXX Handle failure differently ? */ 1099 if (spapr_vio_dma_read(&s->vdev, crq->s.IU_data_ptr, &req->iu, 1100 crq->s.IU_length)) { 1101 fprintf(stderr, "vscsi_got_payload: DMA read failure !\n"); 1102 vscsi_put_req(req); 1103 return; 1104 } 1105 memcpy(&req->crq, crq, sizeof(vscsi_crq)); 1106 1107 if (crq->s.format == VIOSRP_MAD_FORMAT) { 1108 done = vscsi_handle_mad_req(s, req); 1109 } else { 1110 done = vscsi_handle_srp_req(s, req); 1111 } 1112 1113 if (done) { 1114 vscsi_put_req(req); 1115 } 1116 } 1117 1118 1119 static int vscsi_do_crq(struct SpaprVioDevice *dev, uint8_t *crq_data) 1120 { 1121 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); 1122 vscsi_crq crq; 1123 1124 memcpy(crq.raw, crq_data, 16); 1125 crq.s.timeout = be16_to_cpu(crq.s.timeout); 1126 crq.s.IU_length = be16_to_cpu(crq.s.IU_length); 1127 crq.s.IU_data_ptr = be64_to_cpu(crq.s.IU_data_ptr); 1128 1129 trace_spapr_vscsi_do_crq(crq.raw[0], crq.raw[1]); 1130 1131 switch (crq.s.valid) { 1132 case 0xc0: /* Init command/response */ 1133 1134 /* Respond to initialization request */ 1135 if (crq.s.format == 0x01) { 1136 memset(crq.raw, 0, 16); 1137 crq.s.valid = 0xc0; 1138 crq.s.format = 0x02; 1139 spapr_vio_send_crq(dev, crq.raw); 1140 } 1141 1142 /* Note that in hotplug cases, we might get a 0x02 1143 * as a result of us emitting the init request 1144 */ 1145 1146 break; 1147 case 0xff: /* Link event */ 1148 1149 /* Not handled for now */ 1150 1151 break; 1152 case 0x80: /* Payloads */ 1153 switch (crq.s.format) { 1154 case VIOSRP_SRP_FORMAT: /* AKA VSCSI request */ 1155 case VIOSRP_MAD_FORMAT: /* AKA VSCSI response */ 1156 vscsi_got_payload(s, &crq); 1157 break; 1158 case VIOSRP_OS400_FORMAT: 1159 case VIOSRP_AIX_FORMAT: 1160 case VIOSRP_LINUX_FORMAT: 1161 case VIOSRP_INLINE_FORMAT: 1162 fprintf(stderr, "vscsi_do_srq: Unsupported payload format %02x\n", 1163 crq.s.format); 1164 break; 1165 default: 1166 fprintf(stderr, "vscsi_do_srq: Unknown payload format %02x\n", 1167 crq.s.format); 1168 } 1169 break; 1170 default: 1171 fprintf(stderr, "vscsi_do_crq: unknown CRQ %02x %02x ...\n", 1172 crq.raw[0], crq.raw[1]); 1173 }; 1174 1175 return 0; 1176 } 1177 1178 static const struct SCSIBusInfo vscsi_scsi_info = { 1179 .tcq = true, 1180 .max_channel = 7, /* logical unit addressing format */ 1181 .max_target = 63, 1182 .max_lun = 31, 1183 1184 .transfer_data = vscsi_transfer_data, 1185 .complete = vscsi_command_complete, 1186 .cancel = vscsi_request_cancelled, 1187 .save_request = vscsi_save_request, 1188 .load_request = vscsi_load_request, 1189 }; 1190 1191 static void spapr_vscsi_reset(SpaprVioDevice *dev) 1192 { 1193 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); 1194 int i; 1195 1196 memset(s->reqs, 0, sizeof(s->reqs)); 1197 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 1198 s->reqs[i].qtag = i; 1199 } 1200 } 1201 1202 static void spapr_vscsi_realize(SpaprVioDevice *dev, Error **errp) 1203 { 1204 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); 1205 1206 dev->crq.SendFunc = vscsi_do_crq; 1207 1208 scsi_bus_new(&s->bus, sizeof(s->bus), DEVICE(dev), 1209 &vscsi_scsi_info, NULL); 1210 } 1211 1212 void spapr_vscsi_create(SpaprVioBus *bus) 1213 { 1214 DeviceState *dev; 1215 1216 dev = qdev_create(&bus->bus, "spapr-vscsi"); 1217 1218 qdev_init_nofail(dev); 1219 scsi_bus_legacy_handle_cmdline(&VIO_SPAPR_VSCSI_DEVICE(dev)->bus); 1220 } 1221 1222 static int spapr_vscsi_devnode(SpaprVioDevice *dev, void *fdt, int node_off) 1223 { 1224 int ret; 1225 1226 ret = fdt_setprop_cell(fdt, node_off, "#address-cells", 2); 1227 if (ret < 0) { 1228 return ret; 1229 } 1230 1231 ret = fdt_setprop_cell(fdt, node_off, "#size-cells", 0); 1232 if (ret < 0) { 1233 return ret; 1234 } 1235 1236 return 0; 1237 } 1238 1239 static Property spapr_vscsi_properties[] = { 1240 DEFINE_SPAPR_PROPERTIES(VSCSIState, vdev), 1241 DEFINE_PROP_END_OF_LIST(), 1242 }; 1243 1244 static const VMStateDescription vmstate_spapr_vscsi = { 1245 .name = "spapr_vscsi", 1246 .version_id = 1, 1247 .minimum_version_id = 1, 1248 .fields = (VMStateField[]) { 1249 VMSTATE_SPAPR_VIO(vdev, VSCSIState), 1250 /* VSCSI state */ 1251 /* ???? */ 1252 1253 VMSTATE_END_OF_LIST() 1254 }, 1255 }; 1256 1257 static void spapr_vscsi_class_init(ObjectClass *klass, void *data) 1258 { 1259 DeviceClass *dc = DEVICE_CLASS(klass); 1260 SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass); 1261 1262 k->realize = spapr_vscsi_realize; 1263 k->reset = spapr_vscsi_reset; 1264 k->devnode = spapr_vscsi_devnode; 1265 k->dt_name = "v-scsi"; 1266 k->dt_type = "vscsi"; 1267 k->dt_compatible = "IBM,v-scsi"; 1268 k->signal_mask = 0x00000001; 1269 set_bit(DEVICE_CATEGORY_STORAGE, dc->categories); 1270 device_class_set_props(dc, spapr_vscsi_properties); 1271 k->rtce_window_size = 0x10000000; 1272 dc->vmsd = &vmstate_spapr_vscsi; 1273 } 1274 1275 static const TypeInfo spapr_vscsi_info = { 1276 .name = TYPE_VIO_SPAPR_VSCSI_DEVICE, 1277 .parent = TYPE_VIO_SPAPR_DEVICE, 1278 .instance_size = sizeof(VSCSIState), 1279 .class_init = spapr_vscsi_class_init, 1280 }; 1281 1282 static void spapr_vscsi_register_types(void) 1283 { 1284 type_register_static(&spapr_vscsi_info); 1285 } 1286 1287 type_init(spapr_vscsi_register_types) 1288