1 /* 2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator 3 * 4 * PAPR Virtual SCSI, aka ibmvscsi 5 * 6 * Copyright (c) 2010,2011 Benjamin Herrenschmidt, IBM Corporation. 7 * 8 * Permission is hereby granted, free of charge, to any person obtaining a copy 9 * of this software and associated documentation files (the "Software"), to deal 10 * in the Software without restriction, including without limitation the rights 11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell 12 * copies of the Software, and to permit persons to whom the Software is 13 * furnished to do so, subject to the following conditions: 14 * 15 * The above copyright notice and this permission notice shall be included in 16 * all copies or substantial portions of the Software. 17 * 18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR 19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, 20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL 21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER 22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, 23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN 24 * THE SOFTWARE. 25 * 26 * TODO: 27 * 28 * - Cleanups :-) 29 * - Sort out better how to assign devices to VSCSI instances 30 * - Fix residual counts 31 * - Add indirect descriptors support 32 * - Maybe do autosense (PAPR seems to mandate it, linux doesn't care) 33 */ 34 35 #include "qemu/osdep.h" 36 #include "qemu/module.h" 37 #include "cpu.h" 38 #include "hw/hw.h" 39 #include "hw/scsi/scsi.h" 40 #include "migration/vmstate.h" 41 #include "scsi/constants.h" 42 #include "srp.h" 43 #include "hw/qdev.h" 44 #include "hw/ppc/spapr.h" 45 #include "hw/ppc/spapr_vio.h" 46 #include "viosrp.h" 47 #include "trace.h" 48 49 #include <libfdt.h> 50 51 /* 52 * Virtual SCSI device 53 */ 54 55 /* Random numbers */ 56 #define VSCSI_MAX_SECTORS 4096 57 #define VSCSI_REQ_LIMIT 24 58 59 #define SRP_RSP_SENSE_DATA_LEN 18 60 61 #define SRP_REPORT_LUNS_WLUN 0xc10100000000000ULL 62 63 typedef union vscsi_crq { 64 struct viosrp_crq s; 65 uint8_t raw[16]; 66 } vscsi_crq; 67 68 typedef struct vscsi_req { 69 vscsi_crq crq; 70 union viosrp_iu iu; 71 72 /* SCSI request tracking */ 73 SCSIRequest *sreq; 74 uint32_t qtag; /* qemu tag != srp tag */ 75 bool active; 76 bool writing; 77 bool dma_error; 78 uint32_t data_len; 79 uint32_t senselen; 80 uint8_t sense[SCSI_SENSE_BUF_SIZE]; 81 82 /* RDMA related bits */ 83 uint8_t dma_fmt; 84 uint16_t local_desc; 85 uint16_t total_desc; 86 uint16_t cdb_offset; 87 uint16_t cur_desc_num; 88 uint16_t cur_desc_offset; 89 } vscsi_req; 90 91 #define TYPE_VIO_SPAPR_VSCSI_DEVICE "spapr-vscsi" 92 #define VIO_SPAPR_VSCSI_DEVICE(obj) \ 93 OBJECT_CHECK(VSCSIState, (obj), TYPE_VIO_SPAPR_VSCSI_DEVICE) 94 95 typedef struct { 96 SpaprVioDevice vdev; 97 SCSIBus bus; 98 vscsi_req reqs[VSCSI_REQ_LIMIT]; 99 } VSCSIState; 100 101 static struct vscsi_req *vscsi_get_req(VSCSIState *s) 102 { 103 vscsi_req *req; 104 int i; 105 106 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 107 req = &s->reqs[i]; 108 if (!req->active) { 109 memset(req, 0, sizeof(*req)); 110 req->qtag = i; 111 req->active = 1; 112 return req; 113 } 114 } 115 return NULL; 116 } 117 118 static struct vscsi_req *vscsi_find_req(VSCSIState *s, uint64_t srp_tag) 119 { 120 vscsi_req *req; 121 int i; 122 123 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 124 req = &s->reqs[i]; 125 if (req->iu.srp.cmd.tag == srp_tag) { 126 return req; 127 } 128 } 129 return NULL; 130 } 131 132 static void vscsi_put_req(vscsi_req *req) 133 { 134 if (req->sreq != NULL) { 135 scsi_req_unref(req->sreq); 136 } 137 req->sreq = NULL; 138 req->active = 0; 139 } 140 141 static SCSIDevice *vscsi_device_find(SCSIBus *bus, uint64_t srp_lun, int *lun) 142 { 143 int channel = 0, id = 0; 144 145 retry: 146 switch (srp_lun >> 62) { 147 case 0: 148 if ((srp_lun >> 56) != 0) { 149 channel = (srp_lun >> 56) & 0x3f; 150 id = (srp_lun >> 48) & 0xff; 151 srp_lun <<= 16; 152 goto retry; 153 } 154 *lun = (srp_lun >> 48) & 0xff; 155 break; 156 157 case 1: 158 *lun = (srp_lun >> 48) & 0x3fff; 159 break; 160 case 2: 161 channel = (srp_lun >> 53) & 0x7; 162 id = (srp_lun >> 56) & 0x3f; 163 *lun = (srp_lun >> 48) & 0x1f; 164 break; 165 case 3: 166 *lun = -1; 167 return NULL; 168 default: 169 abort(); 170 } 171 172 return scsi_device_find(bus, channel, id, *lun); 173 } 174 175 static int vscsi_send_iu(VSCSIState *s, vscsi_req *req, 176 uint64_t length, uint8_t format) 177 { 178 long rc, rc1; 179 180 /* First copy the SRP */ 181 rc = spapr_vio_dma_write(&s->vdev, req->crq.s.IU_data_ptr, 182 &req->iu, length); 183 if (rc) { 184 fprintf(stderr, "vscsi_send_iu: DMA write failure !\n"); 185 } 186 187 req->crq.s.valid = 0x80; 188 req->crq.s.format = format; 189 req->crq.s.reserved = 0x00; 190 req->crq.s.timeout = cpu_to_be16(0x0000); 191 req->crq.s.IU_length = cpu_to_be16(length); 192 req->crq.s.IU_data_ptr = req->iu.srp.rsp.tag; /* right byte order */ 193 194 if (rc == 0) { 195 req->crq.s.status = VIOSRP_OK; 196 } else { 197 req->crq.s.status = VIOSRP_ADAPTER_FAIL; 198 } 199 200 rc1 = spapr_vio_send_crq(&s->vdev, req->crq.raw); 201 if (rc1) { 202 fprintf(stderr, "vscsi_send_iu: Error sending response\n"); 203 return rc1; 204 } 205 206 return rc; 207 } 208 209 static void vscsi_makeup_sense(VSCSIState *s, vscsi_req *req, 210 uint8_t key, uint8_t asc, uint8_t ascq) 211 { 212 req->senselen = SRP_RSP_SENSE_DATA_LEN; 213 214 /* Valid bit and 'current errors' */ 215 req->sense[0] = (0x1 << 7 | 0x70); 216 /* Sense key */ 217 req->sense[2] = key; 218 /* Additional sense length */ 219 req->sense[7] = 0xa; /* 10 bytes */ 220 /* Additional sense code */ 221 req->sense[12] = asc; 222 req->sense[13] = ascq; 223 } 224 225 static int vscsi_send_rsp(VSCSIState *s, vscsi_req *req, 226 uint8_t status, int32_t res_in, int32_t res_out) 227 { 228 union viosrp_iu *iu = &req->iu; 229 uint64_t tag = iu->srp.rsp.tag; 230 int total_len = sizeof(iu->srp.rsp); 231 uint8_t sol_not = iu->srp.cmd.sol_not; 232 233 trace_spapr_vscsi_send_rsp(status, res_in, res_out); 234 235 memset(iu, 0, sizeof(struct srp_rsp)); 236 iu->srp.rsp.opcode = SRP_RSP; 237 iu->srp.rsp.req_lim_delta = cpu_to_be32(1); 238 iu->srp.rsp.tag = tag; 239 240 /* Handle residuals */ 241 if (res_in < 0) { 242 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIUNDER; 243 res_in = -res_in; 244 } else if (res_in) { 245 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER; 246 } 247 if (res_out < 0) { 248 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOUNDER; 249 res_out = -res_out; 250 } else if (res_out) { 251 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOOVER; 252 } 253 iu->srp.rsp.data_in_res_cnt = cpu_to_be32(res_in); 254 iu->srp.rsp.data_out_res_cnt = cpu_to_be32(res_out); 255 256 /* We don't do response data */ 257 /* iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID; */ 258 iu->srp.rsp.resp_data_len = cpu_to_be32(0); 259 260 /* Handle success vs. failure */ 261 iu->srp.rsp.status = status; 262 if (status) { 263 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2; 264 if (req->senselen) { 265 req->iu.srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID; 266 req->iu.srp.rsp.sense_data_len = cpu_to_be32(req->senselen); 267 memcpy(req->iu.srp.rsp.data, req->sense, req->senselen); 268 total_len += req->senselen; 269 } 270 } else { 271 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1; 272 } 273 274 vscsi_send_iu(s, req, total_len, VIOSRP_SRP_FORMAT); 275 return 0; 276 } 277 278 static inline struct srp_direct_buf vscsi_swap_desc(struct srp_direct_buf desc) 279 { 280 desc.va = be64_to_cpu(desc.va); 281 desc.len = be32_to_cpu(desc.len); 282 return desc; 283 } 284 285 static int vscsi_fetch_desc(VSCSIState *s, struct vscsi_req *req, 286 unsigned n, unsigned buf_offset, 287 struct srp_direct_buf *ret) 288 { 289 struct srp_cmd *cmd = &req->iu.srp.cmd; 290 291 switch (req->dma_fmt) { 292 case SRP_NO_DATA_DESC: { 293 trace_spapr_vscsi_fetch_desc_no_data(); 294 return 0; 295 } 296 case SRP_DATA_DESC_DIRECT: { 297 memcpy(ret, cmd->add_data + req->cdb_offset, sizeof(*ret)); 298 assert(req->cur_desc_num == 0); 299 trace_spapr_vscsi_fetch_desc_direct(); 300 break; 301 } 302 case SRP_DATA_DESC_INDIRECT: { 303 struct srp_indirect_buf *tmp = (struct srp_indirect_buf *) 304 (cmd->add_data + req->cdb_offset); 305 if (n < req->local_desc) { 306 *ret = tmp->desc_list[n]; 307 trace_spapr_vscsi_fetch_desc_indirect(req->qtag, n, 308 req->local_desc); 309 } else if (n < req->total_desc) { 310 int rc; 311 struct srp_direct_buf tbl_desc = vscsi_swap_desc(tmp->table_desc); 312 unsigned desc_offset = n * sizeof(struct srp_direct_buf); 313 314 if (desc_offset >= tbl_desc.len) { 315 trace_spapr_vscsi_fetch_desc_out_of_range(n, desc_offset); 316 return -1; 317 } 318 rc = spapr_vio_dma_read(&s->vdev, tbl_desc.va + desc_offset, 319 ret, sizeof(struct srp_direct_buf)); 320 if (rc) { 321 trace_spapr_vscsi_fetch_desc_dma_read_error(rc); 322 return -1; 323 } 324 trace_spapr_vscsi_fetch_desc_indirect_seg_ext(req->qtag, n, 325 req->total_desc, 326 tbl_desc.va, 327 tbl_desc.len); 328 } else { 329 trace_spapr_vscsi_fetch_desc_out_of_desc(); 330 return 0; 331 } 332 break; 333 } 334 default: 335 fprintf(stderr, "VSCSI: Unknown format %x\n", req->dma_fmt); 336 return -1; 337 } 338 339 *ret = vscsi_swap_desc(*ret); 340 if (buf_offset > ret->len) { 341 trace_spapr_vscsi_fetch_desc_out_of_desc_boundary(buf_offset, 342 req->cur_desc_num, 343 ret->len); 344 return -1; 345 } 346 ret->va += buf_offset; 347 ret->len -= buf_offset; 348 349 trace_spapr_vscsi_fetch_desc_done(req->cur_desc_num, req->cur_desc_offset, 350 ret->va, ret->len); 351 352 return ret->len ? 1 : 0; 353 } 354 355 static int vscsi_srp_direct_data(VSCSIState *s, vscsi_req *req, 356 uint8_t *buf, uint32_t len) 357 { 358 struct srp_direct_buf md; 359 uint32_t llen; 360 int rc = 0; 361 362 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md); 363 if (rc < 0) { 364 return -1; 365 } else if (rc == 0) { 366 return 0; 367 } 368 369 llen = MIN(len, md.len); 370 if (llen) { 371 if (req->writing) { /* writing = to device = reading from memory */ 372 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen); 373 } else { 374 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen); 375 } 376 } 377 378 if (rc) { 379 return -1; 380 } 381 req->cur_desc_offset += llen; 382 383 return llen; 384 } 385 386 static int vscsi_srp_indirect_data(VSCSIState *s, vscsi_req *req, 387 uint8_t *buf, uint32_t len) 388 { 389 struct srp_direct_buf md; 390 int rc = 0; 391 uint32_t llen, total = 0; 392 393 trace_spapr_vscsi_srp_indirect_data(len); 394 395 /* While we have data ... */ 396 while (len) { 397 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md); 398 if (rc < 0) { 399 return -1; 400 } else if (rc == 0) { 401 break; 402 } 403 404 /* Perform transfer */ 405 llen = MIN(len, md.len); 406 if (req->writing) { /* writing = to device = reading from memory */ 407 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen); 408 } else { 409 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen); 410 } 411 if (rc) { 412 trace_spapr_vscsi_srp_indirect_data_rw(req->writing, rc); 413 break; 414 } 415 trace_spapr_vscsi_srp_indirect_data_buf(buf[0], buf[1], buf[2], buf[3]); 416 417 len -= llen; 418 buf += llen; 419 420 total += llen; 421 422 /* Update current position in the current descriptor */ 423 req->cur_desc_offset += llen; 424 if (md.len == llen) { 425 /* Go to the next descriptor if the current one finished */ 426 ++req->cur_desc_num; 427 req->cur_desc_offset = 0; 428 } 429 } 430 431 return rc ? -1 : total; 432 } 433 434 static int vscsi_srp_transfer_data(VSCSIState *s, vscsi_req *req, 435 int writing, uint8_t *buf, uint32_t len) 436 { 437 int err = 0; 438 439 switch (req->dma_fmt) { 440 case SRP_NO_DATA_DESC: 441 trace_spapr_vscsi_srp_transfer_data(len); 442 break; 443 case SRP_DATA_DESC_DIRECT: 444 err = vscsi_srp_direct_data(s, req, buf, len); 445 break; 446 case SRP_DATA_DESC_INDIRECT: 447 err = vscsi_srp_indirect_data(s, req, buf, len); 448 break; 449 } 450 return err; 451 } 452 453 /* Bits from linux srp */ 454 static int data_out_desc_size(struct srp_cmd *cmd) 455 { 456 int size = 0; 457 uint8_t fmt = cmd->buf_fmt >> 4; 458 459 switch (fmt) { 460 case SRP_NO_DATA_DESC: 461 break; 462 case SRP_DATA_DESC_DIRECT: 463 size = sizeof(struct srp_direct_buf); 464 break; 465 case SRP_DATA_DESC_INDIRECT: 466 size = sizeof(struct srp_indirect_buf) + 467 sizeof(struct srp_direct_buf)*cmd->data_out_desc_cnt; 468 break; 469 default: 470 break; 471 } 472 return size; 473 } 474 475 static int vscsi_preprocess_desc(vscsi_req *req) 476 { 477 struct srp_cmd *cmd = &req->iu.srp.cmd; 478 479 req->cdb_offset = cmd->add_cdb_len & ~3; 480 481 if (req->writing) { 482 req->dma_fmt = cmd->buf_fmt >> 4; 483 } else { 484 req->cdb_offset += data_out_desc_size(cmd); 485 req->dma_fmt = cmd->buf_fmt & ((1U << 4) - 1); 486 } 487 488 switch (req->dma_fmt) { 489 case SRP_NO_DATA_DESC: 490 break; 491 case SRP_DATA_DESC_DIRECT: 492 req->total_desc = req->local_desc = 1; 493 break; 494 case SRP_DATA_DESC_INDIRECT: { 495 struct srp_indirect_buf *ind_tmp = (struct srp_indirect_buf *) 496 (cmd->add_data + req->cdb_offset); 497 498 req->total_desc = be32_to_cpu(ind_tmp->table_desc.len) / 499 sizeof(struct srp_direct_buf); 500 req->local_desc = req->writing ? cmd->data_out_desc_cnt : 501 cmd->data_in_desc_cnt; 502 break; 503 } 504 default: 505 fprintf(stderr, 506 "vscsi_preprocess_desc: Unknown format %x\n", req->dma_fmt); 507 return -1; 508 } 509 510 return 0; 511 } 512 513 /* Callback to indicate that the SCSI layer has completed a transfer. */ 514 static void vscsi_transfer_data(SCSIRequest *sreq, uint32_t len) 515 { 516 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); 517 vscsi_req *req = sreq->hba_private; 518 uint8_t *buf; 519 int rc = 0; 520 521 trace_spapr_vscsi_transfer_data(sreq->tag, len, req); 522 if (req == NULL) { 523 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag); 524 return; 525 } 526 527 if (len) { 528 buf = scsi_req_get_buf(sreq); 529 rc = vscsi_srp_transfer_data(s, req, req->writing, buf, len); 530 } 531 if (rc < 0) { 532 fprintf(stderr, "VSCSI: RDMA error rc=%d!\n", rc); 533 req->dma_error = true; 534 scsi_req_cancel(req->sreq); 535 return; 536 } 537 538 /* Start next chunk */ 539 req->data_len -= rc; 540 scsi_req_continue(sreq); 541 } 542 543 /* Callback to indicate that the SCSI layer has completed a transfer. */ 544 static void vscsi_command_complete(SCSIRequest *sreq, uint32_t status, size_t resid) 545 { 546 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); 547 vscsi_req *req = sreq->hba_private; 548 int32_t res_in = 0, res_out = 0; 549 550 trace_spapr_vscsi_command_complete(sreq->tag, status, req); 551 if (req == NULL) { 552 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag); 553 return; 554 } 555 556 if (status == CHECK_CONDITION) { 557 req->senselen = scsi_req_get_sense(req->sreq, req->sense, 558 sizeof(req->sense)); 559 trace_spapr_vscsi_command_complete_sense_data1(req->senselen, 560 req->sense[0], req->sense[1], req->sense[2], req->sense[3], 561 req->sense[4], req->sense[5], req->sense[6], req->sense[7]); 562 trace_spapr_vscsi_command_complete_sense_data2( 563 req->sense[8], req->sense[9], req->sense[10], req->sense[11], 564 req->sense[12], req->sense[13], req->sense[14], req->sense[15]); 565 } 566 567 trace_spapr_vscsi_command_complete_status(status); 568 if (status == 0) { 569 /* We handle overflows, not underflows for normal commands, 570 * but hopefully nobody cares 571 */ 572 if (req->writing) { 573 res_out = req->data_len; 574 } else { 575 res_in = req->data_len; 576 } 577 } 578 vscsi_send_rsp(s, req, status, res_in, res_out); 579 vscsi_put_req(req); 580 } 581 582 static void vscsi_request_cancelled(SCSIRequest *sreq) 583 { 584 vscsi_req *req = sreq->hba_private; 585 586 if (req->dma_error) { 587 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent); 588 589 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); 590 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 591 } 592 vscsi_put_req(req); 593 } 594 595 static const VMStateDescription vmstate_spapr_vscsi_req = { 596 .name = "spapr_vscsi_req", 597 .version_id = 1, 598 .minimum_version_id = 1, 599 .fields = (VMStateField[]) { 600 VMSTATE_BUFFER(crq.raw, vscsi_req), 601 VMSTATE_BUFFER(iu.srp.reserved, vscsi_req), 602 VMSTATE_UINT32(qtag, vscsi_req), 603 VMSTATE_BOOL(active, vscsi_req), 604 VMSTATE_UINT32(data_len, vscsi_req), 605 VMSTATE_BOOL(writing, vscsi_req), 606 VMSTATE_UINT32(senselen, vscsi_req), 607 VMSTATE_BUFFER(sense, vscsi_req), 608 VMSTATE_UINT8(dma_fmt, vscsi_req), 609 VMSTATE_UINT16(local_desc, vscsi_req), 610 VMSTATE_UINT16(total_desc, vscsi_req), 611 VMSTATE_UINT16(cdb_offset, vscsi_req), 612 /*Restart SCSI request from the beginning for now */ 613 /*VMSTATE_UINT16(cur_desc_num, vscsi_req), 614 VMSTATE_UINT16(cur_desc_offset, vscsi_req),*/ 615 VMSTATE_END_OF_LIST() 616 }, 617 }; 618 619 static void vscsi_save_request(QEMUFile *f, SCSIRequest *sreq) 620 { 621 vscsi_req *req = sreq->hba_private; 622 assert(req->active); 623 624 vmstate_save_state(f, &vmstate_spapr_vscsi_req, req, NULL); 625 626 trace_spapr_vscsi_save_request(req->qtag, req->cur_desc_num, 627 req->cur_desc_offset); 628 } 629 630 static void *vscsi_load_request(QEMUFile *f, SCSIRequest *sreq) 631 { 632 SCSIBus *bus = sreq->bus; 633 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(bus->qbus.parent); 634 vscsi_req *req; 635 int rc; 636 637 assert(sreq->tag < VSCSI_REQ_LIMIT); 638 req = &s->reqs[sreq->tag]; 639 assert(!req->active); 640 641 memset(req, 0, sizeof(*req)); 642 rc = vmstate_load_state(f, &vmstate_spapr_vscsi_req, req, 1); 643 if (rc) { 644 fprintf(stderr, "VSCSI: failed loading request tag#%u\n", sreq->tag); 645 return NULL; 646 } 647 assert(req->active); 648 649 req->sreq = scsi_req_ref(sreq); 650 651 trace_spapr_vscsi_load_request(req->qtag, req->cur_desc_num, 652 req->cur_desc_offset); 653 654 return req; 655 } 656 657 static void vscsi_process_login(VSCSIState *s, vscsi_req *req) 658 { 659 union viosrp_iu *iu = &req->iu; 660 struct srp_login_rsp *rsp = &iu->srp.login_rsp; 661 uint64_t tag = iu->srp.rsp.tag; 662 663 trace_spapr_vscsi_process_login(); 664 665 /* TODO handle case that requested size is wrong and 666 * buffer format is wrong 667 */ 668 memset(iu, 0, sizeof(struct srp_login_rsp)); 669 rsp->opcode = SRP_LOGIN_RSP; 670 /* Don't advertise quite as many request as we support to 671 * keep room for management stuff etc... 672 */ 673 rsp->req_lim_delta = cpu_to_be32(VSCSI_REQ_LIMIT-2); 674 rsp->tag = tag; 675 rsp->max_it_iu_len = cpu_to_be32(sizeof(union srp_iu)); 676 rsp->max_ti_iu_len = cpu_to_be32(sizeof(union srp_iu)); 677 /* direct and indirect */ 678 rsp->buf_fmt = cpu_to_be16(SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT); 679 680 vscsi_send_iu(s, req, sizeof(*rsp), VIOSRP_SRP_FORMAT); 681 } 682 683 static void vscsi_inquiry_no_target(VSCSIState *s, vscsi_req *req) 684 { 685 uint8_t *cdb = req->iu.srp.cmd.cdb; 686 uint8_t resp_data[36]; 687 int rc, len, alen; 688 689 /* We don't do EVPD. Also check that page_code is 0 */ 690 if ((cdb[1] & 0x01) || cdb[2] != 0) { 691 /* Send INVALID FIELD IN CDB */ 692 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0); 693 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 694 return; 695 } 696 alen = cdb[3]; 697 alen = (alen << 8) | cdb[4]; 698 len = MIN(alen, 36); 699 700 /* Fake up inquiry using PQ=3 */ 701 memset(resp_data, 0, 36); 702 resp_data[0] = 0x7f; /* Not capable of supporting a device here */ 703 resp_data[2] = 0x06; /* SPS-4 */ 704 resp_data[3] = 0x02; /* Resp data format */ 705 resp_data[4] = 36 - 5; /* Additional length */ 706 resp_data[7] = 0x10; /* Sync transfers */ 707 memcpy(&resp_data[16], "QEMU EMPTY ", 16); 708 memcpy(&resp_data[8], "QEMU ", 8); 709 710 req->writing = 0; 711 vscsi_preprocess_desc(req); 712 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len); 713 if (rc < 0) { 714 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); 715 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 716 } else { 717 vscsi_send_rsp(s, req, 0, 36 - rc, 0); 718 } 719 } 720 721 static void vscsi_report_luns(VSCSIState *s, vscsi_req *req) 722 { 723 BusChild *kid; 724 int i, len, n, rc; 725 uint8_t *resp_data; 726 bool found_lun0; 727 728 n = 0; 729 found_lun0 = false; 730 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) { 731 SCSIDevice *dev = SCSI_DEVICE(kid->child); 732 733 n += 8; 734 if (dev->channel == 0 && dev->id == 0 && dev->lun == 0) { 735 found_lun0 = true; 736 } 737 } 738 if (!found_lun0) { 739 n += 8; 740 } 741 len = n+8; 742 743 resp_data = g_malloc0(len); 744 stl_be_p(resp_data, n); 745 i = found_lun0 ? 8 : 16; 746 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) { 747 DeviceState *qdev = kid->child; 748 SCSIDevice *dev = SCSI_DEVICE(qdev); 749 750 if (dev->id == 0 && dev->channel == 0) { 751 resp_data[i] = 0; /* Use simple LUN for 0 (SAM5 4.7.7.1) */ 752 } else { 753 resp_data[i] = (2 << 6); /* Otherwise LUN addressing (4.7.7.4) */ 754 } 755 resp_data[i] |= dev->id; 756 resp_data[i+1] = (dev->channel << 5); 757 resp_data[i+1] |= dev->lun; 758 i += 8; 759 } 760 761 vscsi_preprocess_desc(req); 762 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len); 763 g_free(resp_data); 764 if (rc < 0) { 765 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0); 766 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 767 } else { 768 vscsi_send_rsp(s, req, 0, len - rc, 0); 769 } 770 } 771 772 static int vscsi_queue_cmd(VSCSIState *s, vscsi_req *req) 773 { 774 union srp_iu *srp = &req->iu.srp; 775 SCSIDevice *sdev; 776 int n, lun; 777 778 if ((srp->cmd.lun == 0 || be64_to_cpu(srp->cmd.lun) == SRP_REPORT_LUNS_WLUN) 779 && srp->cmd.cdb[0] == REPORT_LUNS) { 780 vscsi_report_luns(s, req); 781 return 0; 782 } 783 784 sdev = vscsi_device_find(&s->bus, be64_to_cpu(srp->cmd.lun), &lun); 785 if (!sdev) { 786 trace_spapr_vscsi_queue_cmd_no_drive(be64_to_cpu(srp->cmd.lun)); 787 if (srp->cmd.cdb[0] == INQUIRY) { 788 vscsi_inquiry_no_target(s, req); 789 } else { 790 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0x00); 791 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0); 792 } return 1; 793 } 794 795 req->sreq = scsi_req_new(sdev, req->qtag, lun, srp->cmd.cdb, req); 796 n = scsi_req_enqueue(req->sreq); 797 798 trace_spapr_vscsi_queue_cmd(req->qtag, srp->cmd.cdb[0], 799 scsi_command_name(srp->cmd.cdb[0]), lun, n); 800 801 if (n) { 802 /* Transfer direction must be set before preprocessing the 803 * descriptors 804 */ 805 req->writing = (n < 1); 806 807 /* Preprocess RDMA descriptors */ 808 vscsi_preprocess_desc(req); 809 810 /* Get transfer direction and initiate transfer */ 811 if (n > 0) { 812 req->data_len = n; 813 } else if (n < 0) { 814 req->data_len = -n; 815 } 816 scsi_req_continue(req->sreq); 817 } 818 /* Don't touch req here, it may have been recycled already */ 819 820 return 0; 821 } 822 823 static int vscsi_process_tsk_mgmt(VSCSIState *s, vscsi_req *req) 824 { 825 union viosrp_iu *iu = &req->iu; 826 vscsi_req *tmpreq; 827 int i, lun = 0, resp = SRP_TSK_MGMT_COMPLETE; 828 SCSIDevice *d; 829 uint64_t tag = iu->srp.rsp.tag; 830 uint8_t sol_not = iu->srp.cmd.sol_not; 831 832 fprintf(stderr, "vscsi_process_tsk_mgmt %02x\n", 833 iu->srp.tsk_mgmt.tsk_mgmt_func); 834 835 d = vscsi_device_find(&s->bus, be64_to_cpu(req->iu.srp.tsk_mgmt.lun), &lun); 836 if (!d) { 837 resp = SRP_TSK_MGMT_FIELDS_INVALID; 838 } else { 839 switch (iu->srp.tsk_mgmt.tsk_mgmt_func) { 840 case SRP_TSK_ABORT_TASK: 841 if (d->lun != lun) { 842 resp = SRP_TSK_MGMT_FIELDS_INVALID; 843 break; 844 } 845 846 tmpreq = vscsi_find_req(s, req->iu.srp.tsk_mgmt.task_tag); 847 if (tmpreq && tmpreq->sreq) { 848 assert(tmpreq->sreq->hba_private); 849 scsi_req_cancel(tmpreq->sreq); 850 } 851 break; 852 853 case SRP_TSK_LUN_RESET: 854 if (d->lun != lun) { 855 resp = SRP_TSK_MGMT_FIELDS_INVALID; 856 break; 857 } 858 859 qdev_reset_all(&d->qdev); 860 break; 861 862 case SRP_TSK_ABORT_TASK_SET: 863 case SRP_TSK_CLEAR_TASK_SET: 864 if (d->lun != lun) { 865 resp = SRP_TSK_MGMT_FIELDS_INVALID; 866 break; 867 } 868 869 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 870 tmpreq = &s->reqs[i]; 871 if (tmpreq->iu.srp.cmd.lun != req->iu.srp.tsk_mgmt.lun) { 872 continue; 873 } 874 if (!tmpreq->active || !tmpreq->sreq) { 875 continue; 876 } 877 assert(tmpreq->sreq->hba_private); 878 scsi_req_cancel(tmpreq->sreq); 879 } 880 break; 881 882 case SRP_TSK_CLEAR_ACA: 883 resp = SRP_TSK_MGMT_NOT_SUPPORTED; 884 break; 885 886 default: 887 resp = SRP_TSK_MGMT_FIELDS_INVALID; 888 break; 889 } 890 } 891 892 /* Compose the response here as */ 893 memset(iu, 0, sizeof(struct srp_rsp) + 4); 894 iu->srp.rsp.opcode = SRP_RSP; 895 iu->srp.rsp.req_lim_delta = cpu_to_be32(1); 896 iu->srp.rsp.tag = tag; 897 iu->srp.rsp.flags |= SRP_RSP_FLAG_RSPVALID; 898 iu->srp.rsp.resp_data_len = cpu_to_be32(4); 899 if (resp) { 900 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2; 901 } else { 902 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1; 903 } 904 905 iu->srp.rsp.status = GOOD; 906 iu->srp.rsp.data[3] = resp; 907 908 vscsi_send_iu(s, req, sizeof(iu->srp.rsp) + 4, VIOSRP_SRP_FORMAT); 909 910 return 1; 911 } 912 913 static int vscsi_handle_srp_req(VSCSIState *s, vscsi_req *req) 914 { 915 union srp_iu *srp = &req->iu.srp; 916 int done = 1; 917 uint8_t opcode = srp->rsp.opcode; 918 919 switch (opcode) { 920 case SRP_LOGIN_REQ: 921 vscsi_process_login(s, req); 922 break; 923 case SRP_TSK_MGMT: 924 done = vscsi_process_tsk_mgmt(s, req); 925 break; 926 case SRP_CMD: 927 done = vscsi_queue_cmd(s, req); 928 break; 929 case SRP_LOGIN_RSP: 930 case SRP_I_LOGOUT: 931 case SRP_T_LOGOUT: 932 case SRP_RSP: 933 case SRP_CRED_REQ: 934 case SRP_CRED_RSP: 935 case SRP_AER_REQ: 936 case SRP_AER_RSP: 937 fprintf(stderr, "VSCSI: Unsupported opcode %02x\n", opcode); 938 break; 939 default: 940 fprintf(stderr, "VSCSI: Unknown type %02x\n", opcode); 941 } 942 943 return done; 944 } 945 946 static int vscsi_send_adapter_info(VSCSIState *s, vscsi_req *req) 947 { 948 struct viosrp_adapter_info *sinfo; 949 struct mad_adapter_info_data info; 950 int rc; 951 952 sinfo = &req->iu.mad.adapter_info; 953 954 #if 0 /* What for ? */ 955 rc = spapr_vio_dma_read(&s->vdev, be64_to_cpu(sinfo->buffer), 956 &info, be16_to_cpu(sinfo->common.length)); 957 if (rc) { 958 fprintf(stderr, "vscsi_send_adapter_info: DMA read failure !\n"); 959 } 960 #endif 961 memset(&info, 0, sizeof(info)); 962 strcpy(info.srp_version, SRP_VERSION); 963 memcpy(info.partition_name, "qemu", sizeof("qemu")); 964 info.partition_number = cpu_to_be32(0); 965 info.mad_version = cpu_to_be32(1); 966 info.os_type = cpu_to_be32(2); 967 info.port_max_txu[0] = cpu_to_be32(VSCSI_MAX_SECTORS << 9); 968 969 rc = spapr_vio_dma_write(&s->vdev, be64_to_cpu(sinfo->buffer), 970 &info, be16_to_cpu(sinfo->common.length)); 971 if (rc) { 972 fprintf(stderr, "vscsi_send_adapter_info: DMA write failure !\n"); 973 } 974 975 sinfo->common.status = rc ? cpu_to_be32(1) : 0; 976 977 return vscsi_send_iu(s, req, sizeof(*sinfo), VIOSRP_MAD_FORMAT); 978 } 979 980 static int vscsi_send_capabilities(VSCSIState *s, vscsi_req *req) 981 { 982 struct viosrp_capabilities *vcap; 983 struct capabilities cap = { }; 984 uint16_t len, req_len; 985 uint64_t buffer; 986 int rc; 987 988 vcap = &req->iu.mad.capabilities; 989 req_len = len = be16_to_cpu(vcap->common.length); 990 buffer = be64_to_cpu(vcap->buffer); 991 if (len > sizeof(cap)) { 992 fprintf(stderr, "vscsi_send_capabilities: capabilities size mismatch !\n"); 993 994 /* 995 * Just read and populate the structure that is known. 996 * Zero rest of the structure. 997 */ 998 len = sizeof(cap); 999 } 1000 rc = spapr_vio_dma_read(&s->vdev, buffer, &cap, len); 1001 if (rc) { 1002 fprintf(stderr, "vscsi_send_capabilities: DMA read failure !\n"); 1003 } 1004 1005 /* 1006 * Current implementation does not suppport any migration or 1007 * reservation capabilities. Construct the response telling the 1008 * guest not to use them. 1009 */ 1010 cap.flags = 0; 1011 cap.migration.ecl = 0; 1012 cap.reserve.type = 0; 1013 cap.migration.common.server_support = 0; 1014 cap.reserve.common.server_support = 0; 1015 1016 rc = spapr_vio_dma_write(&s->vdev, buffer, &cap, len); 1017 if (rc) { 1018 fprintf(stderr, "vscsi_send_capabilities: DMA write failure !\n"); 1019 } 1020 if (req_len > len) { 1021 /* 1022 * Being paranoid and lets not worry about the error code 1023 * here. Actual write of the cap is done above. 1024 */ 1025 spapr_vio_dma_set(&s->vdev, (buffer + len), 0, (req_len - len)); 1026 } 1027 vcap->common.status = rc ? cpu_to_be32(1) : 0; 1028 return vscsi_send_iu(s, req, sizeof(*vcap), VIOSRP_MAD_FORMAT); 1029 } 1030 1031 static int vscsi_handle_mad_req(VSCSIState *s, vscsi_req *req) 1032 { 1033 union mad_iu *mad = &req->iu.mad; 1034 bool request_handled = false; 1035 uint64_t retlen = 0; 1036 1037 switch (be32_to_cpu(mad->empty_iu.common.type)) { 1038 case VIOSRP_EMPTY_IU_TYPE: 1039 fprintf(stderr, "Unsupported EMPTY MAD IU\n"); 1040 retlen = sizeof(mad->empty_iu); 1041 break; 1042 case VIOSRP_ERROR_LOG_TYPE: 1043 fprintf(stderr, "Unsupported ERROR LOG MAD IU\n"); 1044 retlen = sizeof(mad->error_log); 1045 break; 1046 case VIOSRP_ADAPTER_INFO_TYPE: 1047 vscsi_send_adapter_info(s, req); 1048 request_handled = true; 1049 break; 1050 case VIOSRP_HOST_CONFIG_TYPE: 1051 retlen = sizeof(mad->host_config); 1052 break; 1053 case VIOSRP_CAPABILITIES_TYPE: 1054 vscsi_send_capabilities(s, req); 1055 request_handled = true; 1056 break; 1057 default: 1058 fprintf(stderr, "VSCSI: Unknown MAD type %02x\n", 1059 be32_to_cpu(mad->empty_iu.common.type)); 1060 /* 1061 * PAPR+ says that "The length field is set to the length 1062 * of the data structure(s) used in the command". 1063 * As we did not recognize the request type, put zero there. 1064 */ 1065 retlen = 0; 1066 } 1067 1068 if (!request_handled) { 1069 mad->empty_iu.common.status = cpu_to_be16(VIOSRP_MAD_NOT_SUPPORTED); 1070 vscsi_send_iu(s, req, retlen, VIOSRP_MAD_FORMAT); 1071 } 1072 1073 return 1; 1074 } 1075 1076 static void vscsi_got_payload(VSCSIState *s, vscsi_crq *crq) 1077 { 1078 vscsi_req *req; 1079 int done; 1080 1081 req = vscsi_get_req(s); 1082 if (req == NULL) { 1083 fprintf(stderr, "VSCSI: Failed to get a request !\n"); 1084 return; 1085 } 1086 1087 /* We only support a limited number of descriptors, we know 1088 * the ibmvscsi driver uses up to 10 max, so it should fit 1089 * in our 256 bytes IUs. If not we'll have to increase the size 1090 * of the structure. 1091 */ 1092 if (crq->s.IU_length > sizeof(union viosrp_iu)) { 1093 fprintf(stderr, "VSCSI: SRP IU too long (%d bytes) !\n", 1094 crq->s.IU_length); 1095 vscsi_put_req(req); 1096 return; 1097 } 1098 1099 /* XXX Handle failure differently ? */ 1100 if (spapr_vio_dma_read(&s->vdev, crq->s.IU_data_ptr, &req->iu, 1101 crq->s.IU_length)) { 1102 fprintf(stderr, "vscsi_got_payload: DMA read failure !\n"); 1103 vscsi_put_req(req); 1104 return; 1105 } 1106 memcpy(&req->crq, crq, sizeof(vscsi_crq)); 1107 1108 if (crq->s.format == VIOSRP_MAD_FORMAT) { 1109 done = vscsi_handle_mad_req(s, req); 1110 } else { 1111 done = vscsi_handle_srp_req(s, req); 1112 } 1113 1114 if (done) { 1115 vscsi_put_req(req); 1116 } 1117 } 1118 1119 1120 static int vscsi_do_crq(struct SpaprVioDevice *dev, uint8_t *crq_data) 1121 { 1122 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); 1123 vscsi_crq crq; 1124 1125 memcpy(crq.raw, crq_data, 16); 1126 crq.s.timeout = be16_to_cpu(crq.s.timeout); 1127 crq.s.IU_length = be16_to_cpu(crq.s.IU_length); 1128 crq.s.IU_data_ptr = be64_to_cpu(crq.s.IU_data_ptr); 1129 1130 trace_spapr_vscsi_do_crq(crq.raw[0], crq.raw[1]); 1131 1132 switch (crq.s.valid) { 1133 case 0xc0: /* Init command/response */ 1134 1135 /* Respond to initialization request */ 1136 if (crq.s.format == 0x01) { 1137 memset(crq.raw, 0, 16); 1138 crq.s.valid = 0xc0; 1139 crq.s.format = 0x02; 1140 spapr_vio_send_crq(dev, crq.raw); 1141 } 1142 1143 /* Note that in hotplug cases, we might get a 0x02 1144 * as a result of us emitting the init request 1145 */ 1146 1147 break; 1148 case 0xff: /* Link event */ 1149 1150 /* Not handled for now */ 1151 1152 break; 1153 case 0x80: /* Payloads */ 1154 switch (crq.s.format) { 1155 case VIOSRP_SRP_FORMAT: /* AKA VSCSI request */ 1156 case VIOSRP_MAD_FORMAT: /* AKA VSCSI response */ 1157 vscsi_got_payload(s, &crq); 1158 break; 1159 case VIOSRP_OS400_FORMAT: 1160 case VIOSRP_AIX_FORMAT: 1161 case VIOSRP_LINUX_FORMAT: 1162 case VIOSRP_INLINE_FORMAT: 1163 fprintf(stderr, "vscsi_do_srq: Unsupported payload format %02x\n", 1164 crq.s.format); 1165 break; 1166 default: 1167 fprintf(stderr, "vscsi_do_srq: Unknown payload format %02x\n", 1168 crq.s.format); 1169 } 1170 break; 1171 default: 1172 fprintf(stderr, "vscsi_do_crq: unknown CRQ %02x %02x ...\n", 1173 crq.raw[0], crq.raw[1]); 1174 }; 1175 1176 return 0; 1177 } 1178 1179 static const struct SCSIBusInfo vscsi_scsi_info = { 1180 .tcq = true, 1181 .max_channel = 7, /* logical unit addressing format */ 1182 .max_target = 63, 1183 .max_lun = 31, 1184 1185 .transfer_data = vscsi_transfer_data, 1186 .complete = vscsi_command_complete, 1187 .cancel = vscsi_request_cancelled, 1188 .save_request = vscsi_save_request, 1189 .load_request = vscsi_load_request, 1190 }; 1191 1192 static void spapr_vscsi_reset(SpaprVioDevice *dev) 1193 { 1194 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); 1195 int i; 1196 1197 memset(s->reqs, 0, sizeof(s->reqs)); 1198 for (i = 0; i < VSCSI_REQ_LIMIT; i++) { 1199 s->reqs[i].qtag = i; 1200 } 1201 } 1202 1203 static void spapr_vscsi_realize(SpaprVioDevice *dev, Error **errp) 1204 { 1205 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev); 1206 1207 dev->crq.SendFunc = vscsi_do_crq; 1208 1209 scsi_bus_new(&s->bus, sizeof(s->bus), DEVICE(dev), 1210 &vscsi_scsi_info, NULL); 1211 } 1212 1213 void spapr_vscsi_create(SpaprVioBus *bus) 1214 { 1215 DeviceState *dev; 1216 1217 dev = qdev_create(&bus->bus, "spapr-vscsi"); 1218 1219 qdev_init_nofail(dev); 1220 scsi_bus_legacy_handle_cmdline(&VIO_SPAPR_VSCSI_DEVICE(dev)->bus); 1221 } 1222 1223 static int spapr_vscsi_devnode(SpaprVioDevice *dev, void *fdt, int node_off) 1224 { 1225 int ret; 1226 1227 ret = fdt_setprop_cell(fdt, node_off, "#address-cells", 2); 1228 if (ret < 0) { 1229 return ret; 1230 } 1231 1232 ret = fdt_setprop_cell(fdt, node_off, "#size-cells", 0); 1233 if (ret < 0) { 1234 return ret; 1235 } 1236 1237 return 0; 1238 } 1239 1240 static Property spapr_vscsi_properties[] = { 1241 DEFINE_SPAPR_PROPERTIES(VSCSIState, vdev), 1242 DEFINE_PROP_END_OF_LIST(), 1243 }; 1244 1245 static const VMStateDescription vmstate_spapr_vscsi = { 1246 .name = "spapr_vscsi", 1247 .version_id = 1, 1248 .minimum_version_id = 1, 1249 .fields = (VMStateField[]) { 1250 VMSTATE_SPAPR_VIO(vdev, VSCSIState), 1251 /* VSCSI state */ 1252 /* ???? */ 1253 1254 VMSTATE_END_OF_LIST() 1255 }, 1256 }; 1257 1258 static void spapr_vscsi_class_init(ObjectClass *klass, void *data) 1259 { 1260 DeviceClass *dc = DEVICE_CLASS(klass); 1261 SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass); 1262 1263 k->realize = spapr_vscsi_realize; 1264 k->reset = spapr_vscsi_reset; 1265 k->devnode = spapr_vscsi_devnode; 1266 k->dt_name = "v-scsi"; 1267 k->dt_type = "vscsi"; 1268 k->dt_compatible = "IBM,v-scsi"; 1269 k->signal_mask = 0x00000001; 1270 set_bit(DEVICE_CATEGORY_STORAGE, dc->categories); 1271 dc->props = spapr_vscsi_properties; 1272 k->rtce_window_size = 0x10000000; 1273 dc->vmsd = &vmstate_spapr_vscsi; 1274 } 1275 1276 static const TypeInfo spapr_vscsi_info = { 1277 .name = TYPE_VIO_SPAPR_VSCSI_DEVICE, 1278 .parent = TYPE_VIO_SPAPR_DEVICE, 1279 .instance_size = sizeof(VSCSIState), 1280 .class_init = spapr_vscsi_class_init, 1281 }; 1282 1283 static void spapr_vscsi_register_types(void) 1284 { 1285 type_register_static(&spapr_vscsi_info); 1286 } 1287 1288 type_init(spapr_vscsi_register_types) 1289