1 /*
2 * QEMU PowerPC pSeries Logical Partition (aka sPAPR) hardware System Emulator
3 *
4 * PAPR Virtual SCSI, aka ibmvscsi
5 *
6 * Copyright (c) 2010,2011 Benjamin Herrenschmidt, IBM Corporation.
7 *
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
14 *
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
17 *
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24 * THE SOFTWARE.
25 *
26 * TODO:
27 *
28 * - Cleanups :-)
29 * - Sort out better how to assign devices to VSCSI instances
30 * - Fix residual counts
31 * - Add indirect descriptors support
32 * - Maybe do autosense (PAPR seems to mandate it, linux doesn't care)
33 */
34
35 #include "qemu/osdep.h"
36 #include "qemu/module.h"
37 #include "hw/scsi/scsi.h"
38 #include "migration/vmstate.h"
39 #include "scsi/constants.h"
40 #include "srp.h"
41 #include "hw/ppc/spapr.h"
42 #include "hw/ppc/spapr_vio.h"
43 #include "hw/qdev-properties.h"
44 #include "viosrp.h"
45 #include "trace.h"
46
47 #include <libfdt.h>
48 #include "qom/object.h"
49
50 /*
51 * Virtual SCSI device
52 */
53
54 /* Random numbers */
55 #define VSCSI_MAX_SECTORS 4096
56 #define VSCSI_REQ_LIMIT 24
57
58 /* Maximum size of a IU payload */
59 #define SRP_MAX_IU_DATA_LEN (SRP_MAX_IU_LEN - sizeof(union srp_iu))
60 #define SRP_RSP_SENSE_DATA_LEN 18
61
62 #define SRP_REPORT_LUNS_WLUN 0xc10100000000000ULL
63
64 typedef union vscsi_crq {
65 struct viosrp_crq s;
66 uint8_t raw[16];
67 } vscsi_crq;
68
69 typedef struct vscsi_req {
70 vscsi_crq crq;
71 uint8_t viosrp_iu_buf[SRP_MAX_IU_LEN];
72
73 /* SCSI request tracking */
74 SCSIRequest *sreq;
75 uint32_t qtag; /* qemu tag != srp tag */
76 bool active;
77 bool writing;
78 bool dma_error;
79 uint32_t data_len;
80 uint32_t senselen;
81 uint8_t sense[SCSI_SENSE_BUF_SIZE];
82
83 /* RDMA related bits */
84 uint8_t dma_fmt;
85 uint16_t local_desc;
86 uint16_t total_desc;
87 uint16_t cdb_offset;
88 uint16_t cur_desc_num;
89 uint16_t cur_desc_offset;
90 } vscsi_req;
91
92 #define TYPE_VIO_SPAPR_VSCSI_DEVICE "spapr-vscsi"
93 OBJECT_DECLARE_SIMPLE_TYPE(VSCSIState, VIO_SPAPR_VSCSI_DEVICE)
94
95 struct VSCSIState {
96 SpaprVioDevice vdev;
97 SCSIBus bus;
98 vscsi_req reqs[VSCSI_REQ_LIMIT];
99 };
100
req_iu(vscsi_req * req)101 static union viosrp_iu *req_iu(vscsi_req *req)
102 {
103 return (union viosrp_iu *)req->viosrp_iu_buf;
104 }
105
vscsi_get_req(VSCSIState * s)106 static struct vscsi_req *vscsi_get_req(VSCSIState *s)
107 {
108 vscsi_req *req;
109 int i;
110
111 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
112 req = &s->reqs[i];
113 if (!req->active) {
114 memset(req, 0, sizeof(*req));
115 req->qtag = i;
116 req->active = 1;
117 return req;
118 }
119 }
120 return NULL;
121 }
122
vscsi_find_req(VSCSIState * s,uint64_t srp_tag)123 static struct vscsi_req *vscsi_find_req(VSCSIState *s, uint64_t srp_tag)
124 {
125 vscsi_req *req;
126 int i;
127
128 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
129 req = &s->reqs[i];
130 if (req_iu(req)->srp.cmd.tag == srp_tag) {
131 return req;
132 }
133 }
134 return NULL;
135 }
136
vscsi_put_req(vscsi_req * req)137 static void vscsi_put_req(vscsi_req *req)
138 {
139 if (req->sreq != NULL) {
140 scsi_req_unref(req->sreq);
141 }
142 req->sreq = NULL;
143 req->active = 0;
144 }
145
vscsi_device_find(SCSIBus * bus,uint64_t srp_lun,int * lun)146 static SCSIDevice *vscsi_device_find(SCSIBus *bus, uint64_t srp_lun, int *lun)
147 {
148 int channel = 0, id = 0;
149
150 retry:
151 switch (srp_lun >> 62) {
152 case 0:
153 if ((srp_lun >> 56) != 0) {
154 channel = (srp_lun >> 56) & 0x3f;
155 id = (srp_lun >> 48) & 0xff;
156 srp_lun <<= 16;
157 goto retry;
158 }
159 *lun = (srp_lun >> 48) & 0xff;
160 break;
161
162 case 1:
163 *lun = (srp_lun >> 48) & 0x3fff;
164 break;
165 case 2:
166 channel = (srp_lun >> 53) & 0x7;
167 id = (srp_lun >> 56) & 0x3f;
168 *lun = (srp_lun >> 48) & 0x1f;
169 break;
170 case 3:
171 *lun = -1;
172 return NULL;
173 default:
174 abort();
175 }
176
177 return scsi_device_find(bus, channel, id, *lun);
178 }
179
vscsi_send_iu(VSCSIState * s,vscsi_req * req,uint64_t length,uint8_t format)180 static int vscsi_send_iu(VSCSIState *s, vscsi_req *req,
181 uint64_t length, uint8_t format)
182 {
183 long rc, rc1;
184
185 assert(length <= SRP_MAX_IU_LEN);
186
187 /* First copy the SRP */
188 rc = spapr_vio_dma_write(&s->vdev, req->crq.s.IU_data_ptr,
189 &req->viosrp_iu_buf, length);
190 if (rc) {
191 fprintf(stderr, "vscsi_send_iu: DMA write failure !\n");
192 }
193
194 req->crq.s.valid = 0x80;
195 req->crq.s.format = format;
196 req->crq.s.reserved = 0x00;
197 req->crq.s.timeout = cpu_to_be16(0x0000);
198 req->crq.s.IU_length = cpu_to_be16(length);
199 req->crq.s.IU_data_ptr = req_iu(req)->srp.rsp.tag; /* right byte order */
200
201 if (rc == 0) {
202 req->crq.s.status = VIOSRP_OK;
203 } else {
204 req->crq.s.status = VIOSRP_ADAPTER_FAIL;
205 }
206
207 rc1 = spapr_vio_send_crq(&s->vdev, req->crq.raw);
208 if (rc1) {
209 fprintf(stderr, "vscsi_send_iu: Error sending response\n");
210 return rc1;
211 }
212
213 return rc;
214 }
215
vscsi_makeup_sense(VSCSIState * s,vscsi_req * req,uint8_t key,uint8_t asc,uint8_t ascq)216 static void vscsi_makeup_sense(VSCSIState *s, vscsi_req *req,
217 uint8_t key, uint8_t asc, uint8_t ascq)
218 {
219 req->senselen = SRP_RSP_SENSE_DATA_LEN;
220
221 /* Valid bit and 'current errors' */
222 req->sense[0] = (0x1 << 7 | 0x70);
223 /* Sense key */
224 req->sense[2] = key;
225 /* Additional sense length */
226 req->sense[7] = 0xa; /* 10 bytes */
227 /* Additional sense code */
228 req->sense[12] = asc;
229 req->sense[13] = ascq;
230 }
231
vscsi_send_rsp(VSCSIState * s,vscsi_req * req,uint8_t status,int32_t res_in,int32_t res_out)232 static int vscsi_send_rsp(VSCSIState *s, vscsi_req *req,
233 uint8_t status, int32_t res_in, int32_t res_out)
234 {
235 union viosrp_iu *iu = req_iu(req);
236 uint64_t tag = iu->srp.rsp.tag;
237 int total_len = sizeof(iu->srp.rsp);
238 uint8_t sol_not = iu->srp.cmd.sol_not;
239
240 trace_spapr_vscsi_send_rsp(status, res_in, res_out);
241
242 memset(iu, 0, sizeof(struct srp_rsp));
243 iu->srp.rsp.opcode = SRP_RSP;
244 iu->srp.rsp.req_lim_delta = cpu_to_be32(1);
245 iu->srp.rsp.tag = tag;
246
247 /* Handle residuals */
248 if (res_in < 0) {
249 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIUNDER;
250 res_in = -res_in;
251 } else if (res_in) {
252 iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER;
253 }
254 if (res_out < 0) {
255 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOUNDER;
256 res_out = -res_out;
257 } else if (res_out) {
258 iu->srp.rsp.flags |= SRP_RSP_FLAG_DOOVER;
259 }
260 iu->srp.rsp.data_in_res_cnt = cpu_to_be32(res_in);
261 iu->srp.rsp.data_out_res_cnt = cpu_to_be32(res_out);
262
263 /* We don't do response data */
264 /* iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID; */
265 iu->srp.rsp.resp_data_len = cpu_to_be32(0);
266
267 /* Handle success vs. failure */
268 iu->srp.rsp.status = status;
269 if (status) {
270 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2;
271 if (req->senselen) {
272 int sense_data_len = MIN(req->senselen, SRP_MAX_IU_DATA_LEN);
273
274 iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
275 iu->srp.rsp.sense_data_len = cpu_to_be32(sense_data_len);
276 memcpy(iu->srp.rsp.data, req->sense, sense_data_len);
277 total_len += sense_data_len;
278 }
279 } else {
280 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1;
281 }
282
283 vscsi_send_iu(s, req, total_len, VIOSRP_SRP_FORMAT);
284 return 0;
285 }
286
vscsi_swap_desc(struct srp_direct_buf desc)287 static inline struct srp_direct_buf vscsi_swap_desc(struct srp_direct_buf desc)
288 {
289 desc.va = be64_to_cpu(desc.va);
290 desc.len = be32_to_cpu(desc.len);
291 return desc;
292 }
293
vscsi_fetch_desc(VSCSIState * s,struct vscsi_req * req,unsigned n,unsigned buf_offset,struct srp_direct_buf * ret)294 static int vscsi_fetch_desc(VSCSIState *s, struct vscsi_req *req,
295 unsigned n, unsigned buf_offset,
296 struct srp_direct_buf *ret)
297 {
298 struct srp_cmd *cmd = &req_iu(req)->srp.cmd;
299
300 switch (req->dma_fmt) {
301 case SRP_NO_DATA_DESC: {
302 trace_spapr_vscsi_fetch_desc_no_data();
303 return 0;
304 }
305 case SRP_DATA_DESC_DIRECT: {
306 memcpy(ret, cmd->add_data + req->cdb_offset, sizeof(*ret));
307 assert(req->cur_desc_num == 0);
308 trace_spapr_vscsi_fetch_desc_direct();
309 break;
310 }
311 case SRP_DATA_DESC_INDIRECT: {
312 struct srp_indirect_buf *tmp = (struct srp_indirect_buf *)
313 (cmd->add_data + req->cdb_offset);
314 if (n < req->local_desc) {
315 *ret = tmp->desc_list[n];
316 trace_spapr_vscsi_fetch_desc_indirect(req->qtag, n,
317 req->local_desc);
318 } else if (n < req->total_desc) {
319 int rc;
320 struct srp_direct_buf tbl_desc = vscsi_swap_desc(tmp->table_desc);
321 unsigned desc_offset = n * sizeof(struct srp_direct_buf);
322
323 if (desc_offset >= tbl_desc.len) {
324 trace_spapr_vscsi_fetch_desc_out_of_range(n, desc_offset);
325 return -1;
326 }
327 rc = spapr_vio_dma_read(&s->vdev, tbl_desc.va + desc_offset,
328 ret, sizeof(struct srp_direct_buf));
329 if (rc) {
330 trace_spapr_vscsi_fetch_desc_dma_read_error(rc);
331 return -1;
332 }
333 trace_spapr_vscsi_fetch_desc_indirect_seg_ext(req->qtag, n,
334 req->total_desc,
335 tbl_desc.va,
336 tbl_desc.len);
337 } else {
338 trace_spapr_vscsi_fetch_desc_out_of_desc();
339 return 0;
340 }
341 break;
342 }
343 default:
344 fprintf(stderr, "VSCSI: Unknown format %x\n", req->dma_fmt);
345 return -1;
346 }
347
348 *ret = vscsi_swap_desc(*ret);
349 if (buf_offset > ret->len) {
350 trace_spapr_vscsi_fetch_desc_out_of_desc_boundary(buf_offset,
351 req->cur_desc_num,
352 ret->len);
353 return -1;
354 }
355 ret->va += buf_offset;
356 ret->len -= buf_offset;
357
358 trace_spapr_vscsi_fetch_desc_done(req->cur_desc_num, req->cur_desc_offset,
359 ret->va, ret->len);
360
361 return ret->len ? 1 : 0;
362 }
363
vscsi_srp_direct_data(VSCSIState * s,vscsi_req * req,uint8_t * buf,uint32_t len)364 static int vscsi_srp_direct_data(VSCSIState *s, vscsi_req *req,
365 uint8_t *buf, uint32_t len)
366 {
367 struct srp_direct_buf md;
368 uint32_t llen;
369 int rc = 0;
370
371 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md);
372 if (rc < 0) {
373 return -1;
374 } else if (rc == 0) {
375 return 0;
376 }
377
378 llen = MIN(len, md.len);
379 if (llen) {
380 if (req->writing) { /* writing = to device = reading from memory */
381 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen);
382 } else {
383 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen);
384 }
385 }
386
387 if (rc) {
388 return -1;
389 }
390 req->cur_desc_offset += llen;
391
392 return llen;
393 }
394
vscsi_srp_indirect_data(VSCSIState * s,vscsi_req * req,uint8_t * buf,uint32_t len)395 static int vscsi_srp_indirect_data(VSCSIState *s, vscsi_req *req,
396 uint8_t *buf, uint32_t len)
397 {
398 struct srp_direct_buf md;
399 int rc = 0;
400 uint32_t llen, total = 0;
401
402 trace_spapr_vscsi_srp_indirect_data(len);
403
404 /* While we have data ... */
405 while (len) {
406 rc = vscsi_fetch_desc(s, req, req->cur_desc_num, req->cur_desc_offset, &md);
407 if (rc < 0) {
408 return -1;
409 } else if (rc == 0) {
410 break;
411 }
412
413 /* Perform transfer */
414 llen = MIN(len, md.len);
415 if (req->writing) { /* writing = to device = reading from memory */
416 rc = spapr_vio_dma_read(&s->vdev, md.va, buf, llen);
417 } else {
418 rc = spapr_vio_dma_write(&s->vdev, md.va, buf, llen);
419 }
420 if (rc) {
421 trace_spapr_vscsi_srp_indirect_data_rw(req->writing, rc);
422 break;
423 }
424 trace_spapr_vscsi_srp_indirect_data_buf(buf[0], buf[1], buf[2], buf[3]);
425
426 len -= llen;
427 buf += llen;
428
429 total += llen;
430
431 /* Update current position in the current descriptor */
432 req->cur_desc_offset += llen;
433 if (md.len == llen) {
434 /* Go to the next descriptor if the current one finished */
435 ++req->cur_desc_num;
436 req->cur_desc_offset = 0;
437 }
438 }
439
440 return rc ? -1 : total;
441 }
442
vscsi_srp_transfer_data(VSCSIState * s,vscsi_req * req,int writing,uint8_t * buf,uint32_t len)443 static int vscsi_srp_transfer_data(VSCSIState *s, vscsi_req *req,
444 int writing, uint8_t *buf, uint32_t len)
445 {
446 int err = 0;
447
448 switch (req->dma_fmt) {
449 case SRP_NO_DATA_DESC:
450 trace_spapr_vscsi_srp_transfer_data(len);
451 break;
452 case SRP_DATA_DESC_DIRECT:
453 err = vscsi_srp_direct_data(s, req, buf, len);
454 break;
455 case SRP_DATA_DESC_INDIRECT:
456 err = vscsi_srp_indirect_data(s, req, buf, len);
457 break;
458 }
459 return err;
460 }
461
462 /* Bits from linux srp */
data_out_desc_size(struct srp_cmd * cmd)463 static int data_out_desc_size(struct srp_cmd *cmd)
464 {
465 int size = 0;
466 uint8_t fmt = cmd->buf_fmt >> 4;
467
468 switch (fmt) {
469 case SRP_NO_DATA_DESC:
470 break;
471 case SRP_DATA_DESC_DIRECT:
472 size = sizeof(struct srp_direct_buf);
473 break;
474 case SRP_DATA_DESC_INDIRECT:
475 size = sizeof(struct srp_indirect_buf) +
476 sizeof(struct srp_direct_buf)*cmd->data_out_desc_cnt;
477 break;
478 default:
479 break;
480 }
481 return size;
482 }
483
vscsi_preprocess_desc(vscsi_req * req)484 static int vscsi_preprocess_desc(vscsi_req *req)
485 {
486 struct srp_cmd *cmd = &req_iu(req)->srp.cmd;
487
488 req->cdb_offset = cmd->add_cdb_len & ~3;
489
490 if (req->writing) {
491 req->dma_fmt = cmd->buf_fmt >> 4;
492 } else {
493 req->cdb_offset += data_out_desc_size(cmd);
494 req->dma_fmt = cmd->buf_fmt & ((1U << 4) - 1);
495 }
496
497 switch (req->dma_fmt) {
498 case SRP_NO_DATA_DESC:
499 break;
500 case SRP_DATA_DESC_DIRECT:
501 req->total_desc = req->local_desc = 1;
502 break;
503 case SRP_DATA_DESC_INDIRECT: {
504 struct srp_indirect_buf *ind_tmp = (struct srp_indirect_buf *)
505 (cmd->add_data + req->cdb_offset);
506
507 req->total_desc = be32_to_cpu(ind_tmp->table_desc.len) /
508 sizeof(struct srp_direct_buf);
509 req->local_desc = req->writing ? cmd->data_out_desc_cnt :
510 cmd->data_in_desc_cnt;
511 break;
512 }
513 default:
514 fprintf(stderr,
515 "vscsi_preprocess_desc: Unknown format %x\n", req->dma_fmt);
516 return -1;
517 }
518
519 return 0;
520 }
521
522 /* Callback to indicate that the SCSI layer has completed a transfer. */
vscsi_transfer_data(SCSIRequest * sreq,uint32_t len)523 static void vscsi_transfer_data(SCSIRequest *sreq, uint32_t len)
524 {
525 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
526 vscsi_req *req = sreq->hba_private;
527 uint8_t *buf;
528 int rc = 0;
529
530 trace_spapr_vscsi_transfer_data(sreq->tag, len, req);
531 if (req == NULL) {
532 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag);
533 return;
534 }
535
536 if (len) {
537 buf = scsi_req_get_buf(sreq);
538 rc = vscsi_srp_transfer_data(s, req, req->writing, buf, len);
539 }
540 if (rc < 0) {
541 fprintf(stderr, "VSCSI: RDMA error rc=%d!\n", rc);
542 req->dma_error = true;
543 scsi_req_cancel(req->sreq);
544 return;
545 }
546
547 /* Start next chunk */
548 req->data_len -= rc;
549 scsi_req_continue(sreq);
550 }
551
552 /* Callback to indicate that the SCSI layer has completed a transfer. */
vscsi_command_complete(SCSIRequest * sreq,size_t resid)553 static void vscsi_command_complete(SCSIRequest *sreq, size_t resid)
554 {
555 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
556 vscsi_req *req = sreq->hba_private;
557 int32_t res_in = 0, res_out = 0;
558
559 trace_spapr_vscsi_command_complete(sreq->tag, sreq->status, req);
560 if (req == NULL) {
561 fprintf(stderr, "VSCSI: Can't find request for tag 0x%x\n", sreq->tag);
562 return;
563 }
564
565 if (sreq->status == CHECK_CONDITION) {
566 req->senselen = scsi_req_get_sense(req->sreq, req->sense,
567 sizeof(req->sense));
568 trace_spapr_vscsi_command_complete_sense_data1(req->senselen,
569 req->sense[0], req->sense[1], req->sense[2], req->sense[3],
570 req->sense[4], req->sense[5], req->sense[6], req->sense[7]);
571 trace_spapr_vscsi_command_complete_sense_data2(
572 req->sense[8], req->sense[9], req->sense[10], req->sense[11],
573 req->sense[12], req->sense[13], req->sense[14], req->sense[15]);
574 }
575
576 trace_spapr_vscsi_command_complete_status(sreq->status);
577 if (sreq->status == 0) {
578 /* We handle overflows, not underflows for normal commands,
579 * but hopefully nobody cares
580 */
581 if (req->writing) {
582 res_out = req->data_len;
583 } else {
584 res_in = req->data_len;
585 }
586 }
587 vscsi_send_rsp(s, req, sreq->status, res_in, res_out);
588 vscsi_put_req(req);
589 }
590
vscsi_request_cancelled(SCSIRequest * sreq)591 static void vscsi_request_cancelled(SCSIRequest *sreq)
592 {
593 vscsi_req *req = sreq->hba_private;
594
595 if (req->dma_error) {
596 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(sreq->bus->qbus.parent);
597
598 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
599 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
600 }
601 vscsi_put_req(req);
602 }
603
604 static const VMStateDescription vmstate_spapr_vscsi_req = {
605 .name = "spapr_vscsi_req",
606 .version_id = 1,
607 .minimum_version_id = 1,
608 .fields = (const VMStateField[]) {
609 VMSTATE_BUFFER(crq.raw, vscsi_req),
610 VMSTATE_BUFFER(viosrp_iu_buf, vscsi_req),
611 VMSTATE_UINT32(qtag, vscsi_req),
612 VMSTATE_BOOL(active, vscsi_req),
613 VMSTATE_UINT32(data_len, vscsi_req),
614 VMSTATE_BOOL(writing, vscsi_req),
615 VMSTATE_UINT32(senselen, vscsi_req),
616 VMSTATE_BUFFER(sense, vscsi_req),
617 VMSTATE_UINT8(dma_fmt, vscsi_req),
618 VMSTATE_UINT16(local_desc, vscsi_req),
619 VMSTATE_UINT16(total_desc, vscsi_req),
620 VMSTATE_UINT16(cdb_offset, vscsi_req),
621 /*Restart SCSI request from the beginning for now */
622 /*VMSTATE_UINT16(cur_desc_num, vscsi_req),
623 VMSTATE_UINT16(cur_desc_offset, vscsi_req),*/
624 VMSTATE_END_OF_LIST()
625 },
626 };
627
vscsi_save_request(QEMUFile * f,SCSIRequest * sreq)628 static void vscsi_save_request(QEMUFile *f, SCSIRequest *sreq)
629 {
630 vscsi_req *req = sreq->hba_private;
631 assert(req->active);
632
633 vmstate_save_state(f, &vmstate_spapr_vscsi_req, req, NULL);
634
635 trace_spapr_vscsi_save_request(req->qtag, req->cur_desc_num,
636 req->cur_desc_offset);
637 }
638
vscsi_load_request(QEMUFile * f,SCSIRequest * sreq)639 static void *vscsi_load_request(QEMUFile *f, SCSIRequest *sreq)
640 {
641 SCSIBus *bus = sreq->bus;
642 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(bus->qbus.parent);
643 vscsi_req *req;
644 int rc;
645
646 assert(sreq->tag < VSCSI_REQ_LIMIT);
647 req = &s->reqs[sreq->tag];
648 assert(!req->active);
649
650 memset(req, 0, sizeof(*req));
651 rc = vmstate_load_state(f, &vmstate_spapr_vscsi_req, req, 1);
652 if (rc) {
653 fprintf(stderr, "VSCSI: failed loading request tag#%u\n", sreq->tag);
654 return NULL;
655 }
656 assert(req->active);
657
658 req->sreq = scsi_req_ref(sreq);
659
660 trace_spapr_vscsi_load_request(req->qtag, req->cur_desc_num,
661 req->cur_desc_offset);
662
663 return req;
664 }
665
vscsi_process_login(VSCSIState * s,vscsi_req * req)666 static void vscsi_process_login(VSCSIState *s, vscsi_req *req)
667 {
668 union viosrp_iu *iu = req_iu(req);
669 struct srp_login_rsp *rsp = &iu->srp.login_rsp;
670 uint64_t tag = iu->srp.rsp.tag;
671
672 trace_spapr_vscsi_process_login();
673
674 /* TODO handle case that requested size is wrong and
675 * buffer format is wrong
676 */
677 memset(iu, 0, sizeof(struct srp_login_rsp));
678 rsp->opcode = SRP_LOGIN_RSP;
679 /* Don't advertise quite as many request as we support to
680 * keep room for management stuff etc...
681 */
682 rsp->req_lim_delta = cpu_to_be32(VSCSI_REQ_LIMIT-2);
683 rsp->tag = tag;
684 rsp->max_it_iu_len = cpu_to_be32(SRP_MAX_IU_LEN);
685 rsp->max_ti_iu_len = cpu_to_be32(SRP_MAX_IU_LEN);
686 /* direct and indirect */
687 rsp->buf_fmt = cpu_to_be16(SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT);
688
689 vscsi_send_iu(s, req, sizeof(*rsp), VIOSRP_SRP_FORMAT);
690 }
691
vscsi_inquiry_no_target(VSCSIState * s,vscsi_req * req)692 static void vscsi_inquiry_no_target(VSCSIState *s, vscsi_req *req)
693 {
694 uint8_t *cdb = req_iu(req)->srp.cmd.cdb;
695 uint8_t resp_data[36];
696 int rc, len, alen;
697
698 /* We don't do EVPD. Also check that page_code is 0 */
699 if ((cdb[1] & 0x01) || cdb[2] != 0) {
700 /* Send INVALID FIELD IN CDB */
701 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0);
702 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
703 return;
704 }
705 alen = cdb[3];
706 alen = (alen << 8) | cdb[4];
707 len = MIN(alen, 36);
708
709 /* Fake up inquiry using PQ=3 */
710 memset(resp_data, 0, 36);
711 resp_data[0] = 0x7f; /* Not capable of supporting a device here */
712 resp_data[2] = 0x06; /* SPS-4 */
713 resp_data[3] = 0x02; /* Resp data format */
714 resp_data[4] = 36 - 5; /* Additional length */
715 resp_data[7] = 0x10; /* Sync transfers */
716 memcpy(&resp_data[16], "QEMU EMPTY ", 16);
717 memcpy(&resp_data[8], "QEMU ", 8);
718
719 req->writing = 0;
720 vscsi_preprocess_desc(req);
721 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len);
722 if (rc < 0) {
723 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
724 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
725 } else {
726 vscsi_send_rsp(s, req, 0, 36 - rc, 0);
727 }
728 }
729
vscsi_report_luns(VSCSIState * s,vscsi_req * req)730 static void vscsi_report_luns(VSCSIState *s, vscsi_req *req)
731 {
732 BusChild *kid;
733 int i, len, n, rc;
734 uint8_t *resp_data;
735 bool found_lun0;
736
737 n = 0;
738 found_lun0 = false;
739 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) {
740 SCSIDevice *dev = SCSI_DEVICE(kid->child);
741
742 n += 8;
743 if (dev->channel == 0 && dev->id == 0 && dev->lun == 0) {
744 found_lun0 = true;
745 }
746 }
747 if (!found_lun0) {
748 n += 8;
749 }
750 len = n+8;
751
752 resp_data = g_malloc0(len);
753 stl_be_p(resp_data, n);
754 i = found_lun0 ? 8 : 16;
755 QTAILQ_FOREACH(kid, &s->bus.qbus.children, sibling) {
756 DeviceState *qdev = kid->child;
757 SCSIDevice *dev = SCSI_DEVICE(qdev);
758
759 if (dev->id == 0 && dev->channel == 0) {
760 resp_data[i] = 0; /* Use simple LUN for 0 (SAM5 4.7.7.1) */
761 } else {
762 resp_data[i] = (2 << 6); /* Otherwise LUN addressing (4.7.7.4) */
763 }
764 resp_data[i] |= dev->id;
765 resp_data[i+1] = (dev->channel << 5);
766 resp_data[i+1] |= dev->lun;
767 i += 8;
768 }
769
770 vscsi_preprocess_desc(req);
771 rc = vscsi_srp_transfer_data(s, req, 0, resp_data, len);
772 g_free(resp_data);
773 if (rc < 0) {
774 vscsi_makeup_sense(s, req, HARDWARE_ERROR, 0, 0);
775 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
776 } else {
777 vscsi_send_rsp(s, req, 0, len - rc, 0);
778 }
779 }
780
vscsi_queue_cmd(VSCSIState * s,vscsi_req * req)781 static int vscsi_queue_cmd(VSCSIState *s, vscsi_req *req)
782 {
783 union srp_iu *srp = &req_iu(req)->srp;
784 SCSIDevice *sdev;
785 int n, lun;
786 size_t cdb_len = sizeof (srp->cmd.cdb) + (srp->cmd.add_cdb_len & ~3);
787
788 if ((srp->cmd.lun == 0 || be64_to_cpu(srp->cmd.lun) == SRP_REPORT_LUNS_WLUN)
789 && srp->cmd.cdb[0] == REPORT_LUNS) {
790 vscsi_report_luns(s, req);
791 return 0;
792 }
793
794 sdev = vscsi_device_find(&s->bus, be64_to_cpu(srp->cmd.lun), &lun);
795 if (!sdev) {
796 trace_spapr_vscsi_queue_cmd_no_drive(be64_to_cpu(srp->cmd.lun));
797 if (srp->cmd.cdb[0] == INQUIRY) {
798 vscsi_inquiry_no_target(s, req);
799 } else {
800 vscsi_makeup_sense(s, req, ILLEGAL_REQUEST, 0x24, 0x00);
801 vscsi_send_rsp(s, req, CHECK_CONDITION, 0, 0);
802 } return 1;
803 }
804
805 req->sreq = scsi_req_new(sdev, req->qtag, lun, srp->cmd.cdb, cdb_len, req);
806 n = scsi_req_enqueue(req->sreq);
807
808 trace_spapr_vscsi_queue_cmd(req->qtag, srp->cmd.cdb[0],
809 scsi_command_name(srp->cmd.cdb[0]), lun, n);
810
811 if (n) {
812 /* Transfer direction must be set before preprocessing the
813 * descriptors
814 */
815 req->writing = (n < 1);
816
817 /* Preprocess RDMA descriptors */
818 vscsi_preprocess_desc(req);
819
820 /* Get transfer direction and initiate transfer */
821 if (n > 0) {
822 req->data_len = n;
823 } else if (n < 0) {
824 req->data_len = -n;
825 }
826 scsi_req_continue(req->sreq);
827 }
828 /* Don't touch req here, it may have been recycled already */
829
830 return 0;
831 }
832
vscsi_process_tsk_mgmt(VSCSIState * s,vscsi_req * req)833 static int vscsi_process_tsk_mgmt(VSCSIState *s, vscsi_req *req)
834 {
835 union viosrp_iu *iu = req_iu(req);
836 vscsi_req *tmpreq;
837 int i, lun = 0, resp = SRP_TSK_MGMT_COMPLETE;
838 SCSIDevice *d;
839 uint64_t tag = iu->srp.rsp.tag;
840 uint8_t sol_not = iu->srp.cmd.sol_not;
841
842 trace_spapr_vscsi_process_tsk_mgmt(iu->srp.tsk_mgmt.tsk_mgmt_func);
843 d = vscsi_device_find(&s->bus,
844 be64_to_cpu(req_iu(req)->srp.tsk_mgmt.lun), &lun);
845 if (!d) {
846 resp = SRP_TSK_MGMT_FIELDS_INVALID;
847 } else {
848 switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
849 case SRP_TSK_ABORT_TASK:
850 if (d->lun != lun) {
851 resp = SRP_TSK_MGMT_FIELDS_INVALID;
852 break;
853 }
854
855 tmpreq = vscsi_find_req(s, req_iu(req)->srp.tsk_mgmt.task_tag);
856 if (tmpreq && tmpreq->sreq) {
857 assert(tmpreq->sreq->hba_private);
858 scsi_req_cancel(tmpreq->sreq);
859 }
860 break;
861
862 case SRP_TSK_LUN_RESET:
863 if (d->lun != lun) {
864 resp = SRP_TSK_MGMT_FIELDS_INVALID;
865 break;
866 }
867
868 device_cold_reset(&d->qdev);
869 break;
870
871 case SRP_TSK_ABORT_TASK_SET:
872 case SRP_TSK_CLEAR_TASK_SET:
873 if (d->lun != lun) {
874 resp = SRP_TSK_MGMT_FIELDS_INVALID;
875 break;
876 }
877
878 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
879 tmpreq = &s->reqs[i];
880 if (req_iu(tmpreq)->srp.cmd.lun
881 != req_iu(req)->srp.tsk_mgmt.lun) {
882 continue;
883 }
884 if (!tmpreq->active || !tmpreq->sreq) {
885 continue;
886 }
887 assert(tmpreq->sreq->hba_private);
888 scsi_req_cancel(tmpreq->sreq);
889 }
890 break;
891
892 case SRP_TSK_CLEAR_ACA:
893 resp = SRP_TSK_MGMT_NOT_SUPPORTED;
894 break;
895
896 default:
897 resp = SRP_TSK_MGMT_FIELDS_INVALID;
898 break;
899 }
900 }
901
902 /* Compose the response here as */
903 QEMU_BUILD_BUG_ON(SRP_MAX_IU_DATA_LEN < 4);
904 memset(iu, 0, sizeof(struct srp_rsp) + 4);
905 iu->srp.rsp.opcode = SRP_RSP;
906 iu->srp.rsp.req_lim_delta = cpu_to_be32(1);
907 iu->srp.rsp.tag = tag;
908 iu->srp.rsp.flags |= SRP_RSP_FLAG_RSPVALID;
909 iu->srp.rsp.resp_data_len = cpu_to_be32(4);
910 if (resp) {
911 iu->srp.rsp.sol_not = (sol_not & 0x04) >> 2;
912 } else {
913 iu->srp.rsp.sol_not = (sol_not & 0x02) >> 1;
914 }
915
916 iu->srp.rsp.status = GOOD;
917 iu->srp.rsp.data[3] = resp;
918
919 vscsi_send_iu(s, req, sizeof(iu->srp.rsp) + 4, VIOSRP_SRP_FORMAT);
920
921 return 1;
922 }
923
vscsi_handle_srp_req(VSCSIState * s,vscsi_req * req)924 static int vscsi_handle_srp_req(VSCSIState *s, vscsi_req *req)
925 {
926 union srp_iu *srp = &req_iu(req)->srp;
927 int done = 1;
928 uint8_t opcode = srp->rsp.opcode;
929
930 switch (opcode) {
931 case SRP_LOGIN_REQ:
932 vscsi_process_login(s, req);
933 break;
934 case SRP_TSK_MGMT:
935 done = vscsi_process_tsk_mgmt(s, req);
936 break;
937 case SRP_CMD:
938 done = vscsi_queue_cmd(s, req);
939 break;
940 case SRP_LOGIN_RSP:
941 case SRP_I_LOGOUT:
942 case SRP_T_LOGOUT:
943 case SRP_RSP:
944 case SRP_CRED_REQ:
945 case SRP_CRED_RSP:
946 case SRP_AER_REQ:
947 case SRP_AER_RSP:
948 fprintf(stderr, "VSCSI: Unsupported opcode %02x\n", opcode);
949 break;
950 default:
951 fprintf(stderr, "VSCSI: Unknown type %02x\n", opcode);
952 }
953
954 return done;
955 }
956
vscsi_send_adapter_info(VSCSIState * s,vscsi_req * req)957 static int vscsi_send_adapter_info(VSCSIState *s, vscsi_req *req)
958 {
959 struct viosrp_adapter_info *sinfo;
960 struct mad_adapter_info_data info;
961 int rc;
962
963 sinfo = &req_iu(req)->mad.adapter_info;
964
965 #if 0 /* What for ? */
966 rc = spapr_vio_dma_read(&s->vdev, be64_to_cpu(sinfo->buffer),
967 &info, be16_to_cpu(sinfo->common.length));
968 if (rc) {
969 fprintf(stderr, "vscsi_send_adapter_info: DMA read failure !\n");
970 }
971 #endif
972 memset(&info, 0, sizeof(info));
973 strcpy(info.srp_version, SRP_VERSION);
974 memcpy(info.partition_name, "qemu", sizeof("qemu"));
975 info.partition_number = cpu_to_be32(0);
976 info.mad_version = cpu_to_be32(1);
977 info.os_type = cpu_to_be32(2);
978 info.port_max_txu[0] = cpu_to_be32(VSCSI_MAX_SECTORS << 9);
979
980 rc = spapr_vio_dma_write(&s->vdev, be64_to_cpu(sinfo->buffer),
981 &info, be16_to_cpu(sinfo->common.length));
982 if (rc) {
983 fprintf(stderr, "vscsi_send_adapter_info: DMA write failure !\n");
984 }
985
986 sinfo->common.status = rc ? cpu_to_be32(1) : 0;
987
988 return vscsi_send_iu(s, req, sizeof(*sinfo), VIOSRP_MAD_FORMAT);
989 }
990
vscsi_send_capabilities(VSCSIState * s,vscsi_req * req)991 static int vscsi_send_capabilities(VSCSIState *s, vscsi_req *req)
992 {
993 struct viosrp_capabilities *vcap;
994 struct capabilities cap = { };
995 uint16_t len, req_len;
996 uint64_t buffer;
997 int rc;
998
999 vcap = &req_iu(req)->mad.capabilities;
1000 req_len = len = be16_to_cpu(vcap->common.length);
1001 buffer = be64_to_cpu(vcap->buffer);
1002 if (len > sizeof(cap)) {
1003 fprintf(stderr, "vscsi_send_capabilities: capabilities size mismatch !\n");
1004
1005 /*
1006 * Just read and populate the structure that is known.
1007 * Zero rest of the structure.
1008 */
1009 len = sizeof(cap);
1010 }
1011 rc = spapr_vio_dma_read(&s->vdev, buffer, &cap, len);
1012 if (rc) {
1013 fprintf(stderr, "vscsi_send_capabilities: DMA read failure !\n");
1014 }
1015
1016 /*
1017 * Current implementation does not support any migration or
1018 * reservation capabilities. Construct the response telling the
1019 * guest not to use them.
1020 */
1021 cap.flags = 0;
1022 cap.migration.ecl = 0;
1023 cap.reserve.type = 0;
1024 cap.migration.common.server_support = 0;
1025 cap.reserve.common.server_support = 0;
1026
1027 rc = spapr_vio_dma_write(&s->vdev, buffer, &cap, len);
1028 if (rc) {
1029 fprintf(stderr, "vscsi_send_capabilities: DMA write failure !\n");
1030 }
1031 if (req_len > len) {
1032 /*
1033 * Being paranoid and lets not worry about the error code
1034 * here. Actual write of the cap is done above.
1035 */
1036 spapr_vio_dma_set(&s->vdev, (buffer + len), 0, (req_len - len));
1037 }
1038 vcap->common.status = rc ? cpu_to_be32(1) : 0;
1039 return vscsi_send_iu(s, req, sizeof(*vcap), VIOSRP_MAD_FORMAT);
1040 }
1041
vscsi_handle_mad_req(VSCSIState * s,vscsi_req * req)1042 static int vscsi_handle_mad_req(VSCSIState *s, vscsi_req *req)
1043 {
1044 union mad_iu *mad = &req_iu(req)->mad;
1045 bool request_handled = false;
1046 uint64_t retlen = 0;
1047
1048 switch (be32_to_cpu(mad->empty_iu.common.type)) {
1049 case VIOSRP_EMPTY_IU_TYPE:
1050 fprintf(stderr, "Unsupported EMPTY MAD IU\n");
1051 retlen = sizeof(mad->empty_iu);
1052 break;
1053 case VIOSRP_ERROR_LOG_TYPE:
1054 fprintf(stderr, "Unsupported ERROR LOG MAD IU\n");
1055 retlen = sizeof(mad->error_log);
1056 break;
1057 case VIOSRP_ADAPTER_INFO_TYPE:
1058 vscsi_send_adapter_info(s, req);
1059 request_handled = true;
1060 break;
1061 case VIOSRP_HOST_CONFIG_TYPE:
1062 retlen = sizeof(mad->host_config);
1063 break;
1064 case VIOSRP_CAPABILITIES_TYPE:
1065 vscsi_send_capabilities(s, req);
1066 request_handled = true;
1067 break;
1068 default:
1069 fprintf(stderr, "VSCSI: Unknown MAD type %02x\n",
1070 be32_to_cpu(mad->empty_iu.common.type));
1071 /*
1072 * PAPR+ says that "The length field is set to the length
1073 * of the data structure(s) used in the command".
1074 * As we did not recognize the request type, put zero there.
1075 */
1076 retlen = 0;
1077 }
1078
1079 if (!request_handled) {
1080 mad->empty_iu.common.status = cpu_to_be16(VIOSRP_MAD_NOT_SUPPORTED);
1081 vscsi_send_iu(s, req, retlen, VIOSRP_MAD_FORMAT);
1082 }
1083
1084 return 1;
1085 }
1086
vscsi_got_payload(VSCSIState * s,vscsi_crq * crq)1087 static void vscsi_got_payload(VSCSIState *s, vscsi_crq *crq)
1088 {
1089 vscsi_req *req;
1090 int done;
1091
1092 req = vscsi_get_req(s);
1093 if (req == NULL) {
1094 fprintf(stderr, "VSCSI: Failed to get a request !\n");
1095 return;
1096 }
1097
1098 /* We only support a limited number of descriptors, we know
1099 * the ibmvscsi driver uses up to 10 max, so it should fit
1100 * in our 256 bytes IUs. If not we'll have to increase the size
1101 * of the structure.
1102 */
1103 if (crq->s.IU_length > SRP_MAX_IU_LEN) {
1104 fprintf(stderr, "VSCSI: SRP IU too long (%d bytes) !\n",
1105 crq->s.IU_length);
1106 vscsi_put_req(req);
1107 return;
1108 }
1109
1110 /* XXX Handle failure differently ? */
1111 if (spapr_vio_dma_read(&s->vdev, crq->s.IU_data_ptr, &req->viosrp_iu_buf,
1112 crq->s.IU_length)) {
1113 fprintf(stderr, "vscsi_got_payload: DMA read failure !\n");
1114 vscsi_put_req(req);
1115 return;
1116 }
1117 memcpy(&req->crq, crq, sizeof(vscsi_crq));
1118
1119 if (crq->s.format == VIOSRP_MAD_FORMAT) {
1120 done = vscsi_handle_mad_req(s, req);
1121 } else {
1122 done = vscsi_handle_srp_req(s, req);
1123 }
1124
1125 if (done) {
1126 vscsi_put_req(req);
1127 }
1128 }
1129
1130
vscsi_do_crq(struct SpaprVioDevice * dev,uint8_t * crq_data)1131 static int vscsi_do_crq(struct SpaprVioDevice *dev, uint8_t *crq_data)
1132 {
1133 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
1134 vscsi_crq crq;
1135
1136 memcpy(crq.raw, crq_data, 16);
1137 crq.s.timeout = be16_to_cpu(crq.s.timeout);
1138 crq.s.IU_length = be16_to_cpu(crq.s.IU_length);
1139 crq.s.IU_data_ptr = be64_to_cpu(crq.s.IU_data_ptr);
1140
1141 trace_spapr_vscsi_do_crq(crq.raw[0], crq.raw[1]);
1142
1143 switch (crq.s.valid) {
1144 case 0xc0: /* Init command/response */
1145
1146 /* Respond to initialization request */
1147 if (crq.s.format == 0x01) {
1148 memset(crq.raw, 0, 16);
1149 crq.s.valid = 0xc0;
1150 crq.s.format = 0x02;
1151 spapr_vio_send_crq(dev, crq.raw);
1152 }
1153
1154 /* Note that in hotplug cases, we might get a 0x02
1155 * as a result of us emitting the init request
1156 */
1157
1158 break;
1159 case 0xff: /* Link event */
1160
1161 /* Not handled for now */
1162
1163 break;
1164 case 0x80: /* Payloads */
1165 switch (crq.s.format) {
1166 case VIOSRP_SRP_FORMAT: /* AKA VSCSI request */
1167 case VIOSRP_MAD_FORMAT: /* AKA VSCSI response */
1168 vscsi_got_payload(s, &crq);
1169 break;
1170 case VIOSRP_OS400_FORMAT:
1171 case VIOSRP_AIX_FORMAT:
1172 case VIOSRP_LINUX_FORMAT:
1173 case VIOSRP_INLINE_FORMAT:
1174 fprintf(stderr, "vscsi_do_srq: Unsupported payload format %02x\n",
1175 crq.s.format);
1176 break;
1177 default:
1178 fprintf(stderr, "vscsi_do_srq: Unknown payload format %02x\n",
1179 crq.s.format);
1180 }
1181 break;
1182 default:
1183 fprintf(stderr, "vscsi_do_crq: unknown CRQ %02x %02x ...\n",
1184 crq.raw[0], crq.raw[1]);
1185 };
1186
1187 return 0;
1188 }
1189
1190 static const struct SCSIBusInfo vscsi_scsi_info = {
1191 .tcq = true,
1192 .max_channel = 7, /* logical unit addressing format */
1193 .max_target = 63,
1194 .max_lun = 31,
1195
1196 .transfer_data = vscsi_transfer_data,
1197 .complete = vscsi_command_complete,
1198 .cancel = vscsi_request_cancelled,
1199 .save_request = vscsi_save_request,
1200 .load_request = vscsi_load_request,
1201 };
1202
spapr_vscsi_reset(SpaprVioDevice * dev)1203 static void spapr_vscsi_reset(SpaprVioDevice *dev)
1204 {
1205 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
1206 int i;
1207
1208 memset(s->reqs, 0, sizeof(s->reqs));
1209 for (i = 0; i < VSCSI_REQ_LIMIT; i++) {
1210 s->reqs[i].qtag = i;
1211 }
1212 }
1213
spapr_vscsi_realize(SpaprVioDevice * dev,Error ** errp)1214 static void spapr_vscsi_realize(SpaprVioDevice *dev, Error **errp)
1215 {
1216 VSCSIState *s = VIO_SPAPR_VSCSI_DEVICE(dev);
1217
1218 dev->crq.SendFunc = vscsi_do_crq;
1219
1220 scsi_bus_init(&s->bus, sizeof(s->bus), DEVICE(dev), &vscsi_scsi_info);
1221
1222 /* ibmvscsi SCSI bus does not allow hotplug. */
1223 qbus_set_hotplug_handler(BUS(&s->bus), NULL);
1224 }
1225
spapr_vscsi_create(SpaprVioBus * bus)1226 void spapr_vscsi_create(SpaprVioBus *bus)
1227 {
1228 DeviceState *dev;
1229
1230 dev = qdev_new("spapr-vscsi");
1231
1232 qdev_realize_and_unref(dev, &bus->bus, &error_fatal);
1233 scsi_bus_legacy_handle_cmdline(&VIO_SPAPR_VSCSI_DEVICE(dev)->bus);
1234 }
1235
spapr_vscsi_devnode(SpaprVioDevice * dev,void * fdt,int node_off)1236 static int spapr_vscsi_devnode(SpaprVioDevice *dev, void *fdt, int node_off)
1237 {
1238 int ret;
1239
1240 ret = fdt_setprop_cell(fdt, node_off, "#address-cells", 2);
1241 if (ret < 0) {
1242 return ret;
1243 }
1244
1245 ret = fdt_setprop_cell(fdt, node_off, "#size-cells", 0);
1246 if (ret < 0) {
1247 return ret;
1248 }
1249
1250 return 0;
1251 }
1252
1253 static Property spapr_vscsi_properties[] = {
1254 DEFINE_SPAPR_PROPERTIES(VSCSIState, vdev),
1255 DEFINE_PROP_END_OF_LIST(),
1256 };
1257
1258 static const VMStateDescription vmstate_spapr_vscsi = {
1259 .name = "spapr_vscsi",
1260 .version_id = 1,
1261 .minimum_version_id = 1,
1262 .fields = (const VMStateField[]) {
1263 VMSTATE_SPAPR_VIO(vdev, VSCSIState),
1264 /* VSCSI state */
1265 /* ???? */
1266
1267 VMSTATE_END_OF_LIST()
1268 },
1269 };
1270
spapr_vscsi_class_init(ObjectClass * klass,void * data)1271 static void spapr_vscsi_class_init(ObjectClass *klass, void *data)
1272 {
1273 DeviceClass *dc = DEVICE_CLASS(klass);
1274 SpaprVioDeviceClass *k = VIO_SPAPR_DEVICE_CLASS(klass);
1275
1276 k->realize = spapr_vscsi_realize;
1277 k->reset = spapr_vscsi_reset;
1278 k->devnode = spapr_vscsi_devnode;
1279 k->dt_name = "v-scsi";
1280 k->dt_type = "vscsi";
1281 k->dt_compatible = "IBM,v-scsi";
1282 k->signal_mask = 0x00000001;
1283 set_bit(DEVICE_CATEGORY_STORAGE, dc->categories);
1284 device_class_set_props(dc, spapr_vscsi_properties);
1285 k->rtce_window_size = 0x10000000;
1286 dc->vmsd = &vmstate_spapr_vscsi;
1287 }
1288
1289 static const TypeInfo spapr_vscsi_info = {
1290 .name = TYPE_VIO_SPAPR_VSCSI_DEVICE,
1291 .parent = TYPE_VIO_SPAPR_DEVICE,
1292 .instance_size = sizeof(VSCSIState),
1293 .class_init = spapr_vscsi_class_init,
1294 };
1295
spapr_vscsi_register_types(void)1296 static void spapr_vscsi_register_types(void)
1297 {
1298 type_register_static(&spapr_vscsi_info);
1299 }
1300
1301 type_init(spapr_vscsi_register_types)
1302