xref: /openbmc/qemu/hw/ide/ahci.c (revision 21f5826a)
1 /*
2  * QEMU AHCI Emulation
3  *
4  * Copyright (c) 2010 qiaochong@loongson.cn
5  * Copyright (c) 2010 Roland Elek <elek.roland@gmail.com>
6  * Copyright (c) 2010 Sebastian Herbszt <herbszt@gmx.de>
7  * Copyright (c) 2010 Alexander Graf <agraf@suse.de>
8  *
9  * This library is free software; you can redistribute it and/or
10  * modify it under the terms of the GNU Lesser General Public
11  * License as published by the Free Software Foundation; either
12  * version 2 of the License, or (at your option) any later version.
13  *
14  * This library is distributed in the hope that it will be useful,
15  * but WITHOUT ANY WARRANTY; without even the implied warranty of
16  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
17  * Lesser General Public License for more details.
18  *
19  * You should have received a copy of the GNU Lesser General Public
20  * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21  *
22  */
23 
24 #include <hw/hw.h>
25 #include <hw/pci/msi.h>
26 #include <hw/i386/pc.h>
27 #include <hw/pci/pci.h>
28 #include <hw/sysbus.h>
29 
30 #include "monitor/monitor.h"
31 #include "sysemu/block-backend.h"
32 #include "sysemu/dma.h"
33 #include "internal.h"
34 #include <hw/ide/pci.h>
35 #include <hw/ide/ahci.h>
36 
37 #define DEBUG_AHCI 0
38 
39 #define DPRINTF(port, fmt, ...) \
40 do { \
41     if (DEBUG_AHCI) { \
42         fprintf(stderr, "ahci: %s: [%d] ", __func__, port); \
43         fprintf(stderr, fmt, ## __VA_ARGS__); \
44     } \
45 } while (0)
46 
47 static void check_cmd(AHCIState *s, int port);
48 static int handle_cmd(AHCIState *s,int port,int slot);
49 static void ahci_reset_port(AHCIState *s, int port);
50 static void ahci_write_fis_d2h(AHCIDevice *ad, uint8_t *cmd_fis);
51 static void ahci_init_d2h(AHCIDevice *ad);
52 static int ahci_dma_prepare_buf(IDEDMA *dma, int is_write);
53 static void ahci_commit_buf(IDEDMA *dma, uint32_t tx_bytes);
54 
55 
56 static uint32_t  ahci_port_read(AHCIState *s, int port, int offset)
57 {
58     uint32_t val;
59     AHCIPortRegs *pr;
60     pr = &s->dev[port].port_regs;
61 
62     switch (offset) {
63     case PORT_LST_ADDR:
64         val = pr->lst_addr;
65         break;
66     case PORT_LST_ADDR_HI:
67         val = pr->lst_addr_hi;
68         break;
69     case PORT_FIS_ADDR:
70         val = pr->fis_addr;
71         break;
72     case PORT_FIS_ADDR_HI:
73         val = pr->fis_addr_hi;
74         break;
75     case PORT_IRQ_STAT:
76         val = pr->irq_stat;
77         break;
78     case PORT_IRQ_MASK:
79         val = pr->irq_mask;
80         break;
81     case PORT_CMD:
82         val = pr->cmd;
83         break;
84     case PORT_TFDATA:
85         val = pr->tfdata;
86         break;
87     case PORT_SIG:
88         val = pr->sig;
89         break;
90     case PORT_SCR_STAT:
91         if (s->dev[port].port.ifs[0].blk) {
92             val = SATA_SCR_SSTATUS_DET_DEV_PRESENT_PHY_UP |
93                   SATA_SCR_SSTATUS_SPD_GEN1 | SATA_SCR_SSTATUS_IPM_ACTIVE;
94         } else {
95             val = SATA_SCR_SSTATUS_DET_NODEV;
96         }
97         break;
98     case PORT_SCR_CTL:
99         val = pr->scr_ctl;
100         break;
101     case PORT_SCR_ERR:
102         val = pr->scr_err;
103         break;
104     case PORT_SCR_ACT:
105         pr->scr_act &= ~s->dev[port].finished;
106         s->dev[port].finished = 0;
107         val = pr->scr_act;
108         break;
109     case PORT_CMD_ISSUE:
110         val = pr->cmd_issue;
111         break;
112     case PORT_RESERVED:
113     default:
114         val = 0;
115     }
116     DPRINTF(port, "offset: 0x%x val: 0x%x\n", offset, val);
117     return val;
118 
119 }
120 
121 static void ahci_irq_raise(AHCIState *s, AHCIDevice *dev)
122 {
123     AHCIPCIState *d = container_of(s, AHCIPCIState, ahci);
124     PCIDevice *pci_dev =
125         (PCIDevice *)object_dynamic_cast(OBJECT(d), TYPE_PCI_DEVICE);
126 
127     DPRINTF(0, "raise irq\n");
128 
129     if (pci_dev && msi_enabled(pci_dev)) {
130         msi_notify(pci_dev, 0);
131     } else {
132         qemu_irq_raise(s->irq);
133     }
134 }
135 
136 static void ahci_irq_lower(AHCIState *s, AHCIDevice *dev)
137 {
138     AHCIPCIState *d = container_of(s, AHCIPCIState, ahci);
139     PCIDevice *pci_dev =
140         (PCIDevice *)object_dynamic_cast(OBJECT(d), TYPE_PCI_DEVICE);
141 
142     DPRINTF(0, "lower irq\n");
143 
144     if (!pci_dev || !msi_enabled(pci_dev)) {
145         qemu_irq_lower(s->irq);
146     }
147 }
148 
149 static void ahci_check_irq(AHCIState *s)
150 {
151     int i;
152 
153     DPRINTF(-1, "check irq %#x\n", s->control_regs.irqstatus);
154 
155     s->control_regs.irqstatus = 0;
156     for (i = 0; i < s->ports; i++) {
157         AHCIPortRegs *pr = &s->dev[i].port_regs;
158         if (pr->irq_stat & pr->irq_mask) {
159             s->control_regs.irqstatus |= (1 << i);
160         }
161     }
162 
163     if (s->control_regs.irqstatus &&
164         (s->control_regs.ghc & HOST_CTL_IRQ_EN)) {
165             ahci_irq_raise(s, NULL);
166     } else {
167         ahci_irq_lower(s, NULL);
168     }
169 }
170 
171 static void ahci_trigger_irq(AHCIState *s, AHCIDevice *d,
172                              int irq_type)
173 {
174     DPRINTF(d->port_no, "trigger irq %#x -> %x\n",
175             irq_type, d->port_regs.irq_mask & irq_type);
176 
177     d->port_regs.irq_stat |= irq_type;
178     ahci_check_irq(s);
179 }
180 
181 static void map_page(AddressSpace *as, uint8_t **ptr, uint64_t addr,
182                      uint32_t wanted)
183 {
184     hwaddr len = wanted;
185 
186     if (*ptr) {
187         dma_memory_unmap(as, *ptr, len, DMA_DIRECTION_FROM_DEVICE, len);
188     }
189 
190     *ptr = dma_memory_map(as, addr, &len, DMA_DIRECTION_FROM_DEVICE);
191     if (len < wanted) {
192         dma_memory_unmap(as, *ptr, len, DMA_DIRECTION_FROM_DEVICE, len);
193         *ptr = NULL;
194     }
195 }
196 
197 static void  ahci_port_write(AHCIState *s, int port, int offset, uint32_t val)
198 {
199     AHCIPortRegs *pr = &s->dev[port].port_regs;
200 
201     DPRINTF(port, "offset: 0x%x val: 0x%x\n", offset, val);
202     switch (offset) {
203         case PORT_LST_ADDR:
204             pr->lst_addr = val;
205             map_page(s->as, &s->dev[port].lst,
206                      ((uint64_t)pr->lst_addr_hi << 32) | pr->lst_addr, 1024);
207             s->dev[port].cur_cmd = NULL;
208             break;
209         case PORT_LST_ADDR_HI:
210             pr->lst_addr_hi = val;
211             map_page(s->as, &s->dev[port].lst,
212                      ((uint64_t)pr->lst_addr_hi << 32) | pr->lst_addr, 1024);
213             s->dev[port].cur_cmd = NULL;
214             break;
215         case PORT_FIS_ADDR:
216             pr->fis_addr = val;
217             map_page(s->as, &s->dev[port].res_fis,
218                      ((uint64_t)pr->fis_addr_hi << 32) | pr->fis_addr, 256);
219             break;
220         case PORT_FIS_ADDR_HI:
221             pr->fis_addr_hi = val;
222             map_page(s->as, &s->dev[port].res_fis,
223                      ((uint64_t)pr->fis_addr_hi << 32) | pr->fis_addr, 256);
224             break;
225         case PORT_IRQ_STAT:
226             pr->irq_stat &= ~val;
227             ahci_check_irq(s);
228             break;
229         case PORT_IRQ_MASK:
230             pr->irq_mask = val & 0xfdc000ff;
231             ahci_check_irq(s);
232             break;
233         case PORT_CMD:
234             pr->cmd = val & ~(PORT_CMD_LIST_ON | PORT_CMD_FIS_ON);
235 
236             if (pr->cmd & PORT_CMD_START) {
237                 pr->cmd |= PORT_CMD_LIST_ON;
238             }
239 
240             if (pr->cmd & PORT_CMD_FIS_RX) {
241                 pr->cmd |= PORT_CMD_FIS_ON;
242             }
243 
244             /* XXX usually the FIS would be pending on the bus here and
245                    issuing deferred until the OS enables FIS receival.
246                    Instead, we only submit it once - which works in most
247                    cases, but is a hack. */
248             if ((pr->cmd & PORT_CMD_FIS_ON) &&
249                 !s->dev[port].init_d2h_sent) {
250                 ahci_init_d2h(&s->dev[port]);
251                 s->dev[port].init_d2h_sent = true;
252             }
253 
254             check_cmd(s, port);
255             break;
256         case PORT_TFDATA:
257             /* Read Only. */
258             break;
259         case PORT_SIG:
260             /* Read Only */
261             break;
262         case PORT_SCR_STAT:
263             /* Read Only */
264             break;
265         case PORT_SCR_CTL:
266             if (((pr->scr_ctl & AHCI_SCR_SCTL_DET) == 1) &&
267                 ((val & AHCI_SCR_SCTL_DET) == 0)) {
268                 ahci_reset_port(s, port);
269             }
270             pr->scr_ctl = val;
271             break;
272         case PORT_SCR_ERR:
273             pr->scr_err &= ~val;
274             break;
275         case PORT_SCR_ACT:
276             /* RW1 */
277             pr->scr_act |= val;
278             break;
279         case PORT_CMD_ISSUE:
280             pr->cmd_issue |= val;
281             check_cmd(s, port);
282             break;
283         default:
284             break;
285     }
286 }
287 
288 static uint64_t ahci_mem_read(void *opaque, hwaddr addr,
289                               unsigned size)
290 {
291     AHCIState *s = opaque;
292     uint32_t val = 0;
293 
294     if (addr < AHCI_GENERIC_HOST_CONTROL_REGS_MAX_ADDR) {
295         switch (addr) {
296         case HOST_CAP:
297             val = s->control_regs.cap;
298             break;
299         case HOST_CTL:
300             val = s->control_regs.ghc;
301             break;
302         case HOST_IRQ_STAT:
303             val = s->control_regs.irqstatus;
304             break;
305         case HOST_PORTS_IMPL:
306             val = s->control_regs.impl;
307             break;
308         case HOST_VERSION:
309             val = s->control_regs.version;
310             break;
311         }
312 
313         DPRINTF(-1, "(addr 0x%08X), val 0x%08X\n", (unsigned) addr, val);
314     } else if ((addr >= AHCI_PORT_REGS_START_ADDR) &&
315                (addr < (AHCI_PORT_REGS_START_ADDR +
316                 (s->ports * AHCI_PORT_ADDR_OFFSET_LEN)))) {
317         val = ahci_port_read(s, (addr - AHCI_PORT_REGS_START_ADDR) >> 7,
318                              addr & AHCI_PORT_ADDR_OFFSET_MASK);
319     }
320 
321     return val;
322 }
323 
324 
325 
326 static void ahci_mem_write(void *opaque, hwaddr addr,
327                            uint64_t val, unsigned size)
328 {
329     AHCIState *s = opaque;
330 
331     /* Only aligned reads are allowed on AHCI */
332     if (addr & 3) {
333         fprintf(stderr, "ahci: Mis-aligned write to addr 0x"
334                 TARGET_FMT_plx "\n", addr);
335         return;
336     }
337 
338     if (addr < AHCI_GENERIC_HOST_CONTROL_REGS_MAX_ADDR) {
339         DPRINTF(-1, "(addr 0x%08X), val 0x%08"PRIX64"\n", (unsigned) addr, val);
340 
341         switch (addr) {
342             case HOST_CAP: /* R/WO, RO */
343                 /* FIXME handle R/WO */
344                 break;
345             case HOST_CTL: /* R/W */
346                 if (val & HOST_CTL_RESET) {
347                     DPRINTF(-1, "HBA Reset\n");
348                     ahci_reset(s);
349                 } else {
350                     s->control_regs.ghc = (val & 0x3) | HOST_CTL_AHCI_EN;
351                     ahci_check_irq(s);
352                 }
353                 break;
354             case HOST_IRQ_STAT: /* R/WC, RO */
355                 s->control_regs.irqstatus &= ~val;
356                 ahci_check_irq(s);
357                 break;
358             case HOST_PORTS_IMPL: /* R/WO, RO */
359                 /* FIXME handle R/WO */
360                 break;
361             case HOST_VERSION: /* RO */
362                 /* FIXME report write? */
363                 break;
364             default:
365                 DPRINTF(-1, "write to unknown register 0x%x\n", (unsigned)addr);
366         }
367     } else if ((addr >= AHCI_PORT_REGS_START_ADDR) &&
368                (addr < (AHCI_PORT_REGS_START_ADDR +
369                 (s->ports * AHCI_PORT_ADDR_OFFSET_LEN)))) {
370         ahci_port_write(s, (addr - AHCI_PORT_REGS_START_ADDR) >> 7,
371                         addr & AHCI_PORT_ADDR_OFFSET_MASK, val);
372     }
373 
374 }
375 
376 static const MemoryRegionOps ahci_mem_ops = {
377     .read = ahci_mem_read,
378     .write = ahci_mem_write,
379     .endianness = DEVICE_LITTLE_ENDIAN,
380 };
381 
382 static uint64_t ahci_idp_read(void *opaque, hwaddr addr,
383                               unsigned size)
384 {
385     AHCIState *s = opaque;
386 
387     if (addr == s->idp_offset) {
388         /* index register */
389         return s->idp_index;
390     } else if (addr == s->idp_offset + 4) {
391         /* data register - do memory read at location selected by index */
392         return ahci_mem_read(opaque, s->idp_index, size);
393     } else {
394         return 0;
395     }
396 }
397 
398 static void ahci_idp_write(void *opaque, hwaddr addr,
399                            uint64_t val, unsigned size)
400 {
401     AHCIState *s = opaque;
402 
403     if (addr == s->idp_offset) {
404         /* index register - mask off reserved bits */
405         s->idp_index = (uint32_t)val & ((AHCI_MEM_BAR_SIZE - 1) & ~3);
406     } else if (addr == s->idp_offset + 4) {
407         /* data register - do memory write at location selected by index */
408         ahci_mem_write(opaque, s->idp_index, val, size);
409     }
410 }
411 
412 static const MemoryRegionOps ahci_idp_ops = {
413     .read = ahci_idp_read,
414     .write = ahci_idp_write,
415     .endianness = DEVICE_LITTLE_ENDIAN,
416 };
417 
418 
419 static void ahci_reg_init(AHCIState *s)
420 {
421     int i;
422 
423     s->control_regs.cap = (s->ports - 1) |
424                           (AHCI_NUM_COMMAND_SLOTS << 8) |
425                           (AHCI_SUPPORTED_SPEED_GEN1 << AHCI_SUPPORTED_SPEED) |
426                           HOST_CAP_NCQ | HOST_CAP_AHCI;
427 
428     s->control_regs.impl = (1 << s->ports) - 1;
429 
430     s->control_regs.version = AHCI_VERSION_1_0;
431 
432     for (i = 0; i < s->ports; i++) {
433         s->dev[i].port_state = STATE_RUN;
434     }
435 }
436 
437 static void check_cmd(AHCIState *s, int port)
438 {
439     AHCIPortRegs *pr = &s->dev[port].port_regs;
440     int slot;
441 
442     if ((pr->cmd & PORT_CMD_START) && pr->cmd_issue) {
443         for (slot = 0; (slot < 32) && pr->cmd_issue; slot++) {
444             if ((pr->cmd_issue & (1U << slot)) &&
445                 !handle_cmd(s, port, slot)) {
446                 pr->cmd_issue &= ~(1U << slot);
447             }
448         }
449     }
450 }
451 
452 static void ahci_check_cmd_bh(void *opaque)
453 {
454     AHCIDevice *ad = opaque;
455 
456     qemu_bh_delete(ad->check_bh);
457     ad->check_bh = NULL;
458 
459     if ((ad->busy_slot != -1) &&
460         !(ad->port.ifs[0].status & (BUSY_STAT|DRQ_STAT))) {
461         /* no longer busy */
462         ad->port_regs.cmd_issue &= ~(1 << ad->busy_slot);
463         ad->busy_slot = -1;
464     }
465 
466     check_cmd(ad->hba, ad->port_no);
467 }
468 
469 static void ahci_init_d2h(AHCIDevice *ad)
470 {
471     uint8_t init_fis[20];
472     IDEState *ide_state = &ad->port.ifs[0];
473 
474     memset(init_fis, 0, sizeof(init_fis));
475 
476     init_fis[4] = 1;
477     init_fis[12] = 1;
478 
479     if (ide_state->drive_kind == IDE_CD) {
480         init_fis[5] = ide_state->lcyl;
481         init_fis[6] = ide_state->hcyl;
482     }
483 
484     ahci_write_fis_d2h(ad, init_fis);
485 }
486 
487 static void ahci_reset_port(AHCIState *s, int port)
488 {
489     AHCIDevice *d = &s->dev[port];
490     AHCIPortRegs *pr = &d->port_regs;
491     IDEState *ide_state = &d->port.ifs[0];
492     int i;
493 
494     DPRINTF(port, "reset port\n");
495 
496     ide_bus_reset(&d->port);
497     ide_state->ncq_queues = AHCI_MAX_CMDS;
498 
499     pr->scr_stat = 0;
500     pr->scr_err = 0;
501     pr->scr_act = 0;
502     pr->tfdata = 0x7F;
503     pr->sig = 0xFFFFFFFF;
504     d->busy_slot = -1;
505     d->init_d2h_sent = false;
506 
507     ide_state = &s->dev[port].port.ifs[0];
508     if (!ide_state->blk) {
509         return;
510     }
511 
512     /* reset ncq queue */
513     for (i = 0; i < AHCI_MAX_CMDS; i++) {
514         NCQTransferState *ncq_tfs = &s->dev[port].ncq_tfs[i];
515         if (!ncq_tfs->used) {
516             continue;
517         }
518 
519         if (ncq_tfs->aiocb) {
520             blk_aio_cancel(ncq_tfs->aiocb);
521             ncq_tfs->aiocb = NULL;
522         }
523 
524         /* Maybe we just finished the request thanks to blk_aio_cancel() */
525         if (!ncq_tfs->used) {
526             continue;
527         }
528 
529         qemu_sglist_destroy(&ncq_tfs->sglist);
530         ncq_tfs->used = 0;
531     }
532 
533     s->dev[port].port_state = STATE_RUN;
534     if (!ide_state->blk) {
535         pr->sig = 0;
536         ide_state->status = SEEK_STAT | WRERR_STAT;
537     } else if (ide_state->drive_kind == IDE_CD) {
538         pr->sig = SATA_SIGNATURE_CDROM;
539         ide_state->lcyl = 0x14;
540         ide_state->hcyl = 0xeb;
541         DPRINTF(port, "set lcyl = %d\n", ide_state->lcyl);
542         ide_state->status = SEEK_STAT | WRERR_STAT | READY_STAT;
543     } else {
544         pr->sig = SATA_SIGNATURE_DISK;
545         ide_state->status = SEEK_STAT | WRERR_STAT;
546     }
547 
548     ide_state->error = 1;
549     ahci_init_d2h(d);
550 }
551 
552 static void debug_print_fis(uint8_t *fis, int cmd_len)
553 {
554 #if DEBUG_AHCI
555     int i;
556 
557     fprintf(stderr, "fis:");
558     for (i = 0; i < cmd_len; i++) {
559         if ((i & 0xf) == 0) {
560             fprintf(stderr, "\n%02x:",i);
561         }
562         fprintf(stderr, "%02x ",fis[i]);
563     }
564     fprintf(stderr, "\n");
565 #endif
566 }
567 
568 static void ahci_write_fis_sdb(AHCIState *s, int port, uint32_t finished)
569 {
570     AHCIDevice *ad = &s->dev[port];
571     AHCIPortRegs *pr = &ad->port_regs;
572     IDEState *ide_state;
573     SDBFIS *sdb_fis;
574 
575     if (!s->dev[port].res_fis ||
576         !(pr->cmd & PORT_CMD_FIS_RX)) {
577         return;
578     }
579 
580     sdb_fis = (SDBFIS *)&ad->res_fis[RES_FIS_SDBFIS];
581     ide_state = &ad->port.ifs[0];
582 
583     sdb_fis->type = SATA_FIS_TYPE_SDB;
584     /* Interrupt pending & Notification bit */
585     sdb_fis->flags = (ad->hba->control_regs.irqstatus ? (1 << 6) : 0);
586     sdb_fis->status = ide_state->status & 0x77;
587     sdb_fis->error = ide_state->error;
588     /* update SAct field in SDB_FIS */
589     s->dev[port].finished |= finished;
590     sdb_fis->payload = cpu_to_le32(ad->finished);
591 
592     /* Update shadow registers (except BSY 0x80 and DRQ 0x08) */
593     pr->tfdata = (ad->port.ifs[0].error << 8) |
594         (ad->port.ifs[0].status & 0x77) |
595         (pr->tfdata & 0x88);
596 
597     ahci_trigger_irq(s, ad, PORT_IRQ_SDB_FIS);
598 }
599 
600 static void ahci_write_fis_pio(AHCIDevice *ad, uint16_t len)
601 {
602     AHCIPortRegs *pr = &ad->port_regs;
603     uint8_t *pio_fis, *cmd_fis;
604     uint64_t tbl_addr;
605     dma_addr_t cmd_len = 0x80;
606     IDEState *s = &ad->port.ifs[0];
607 
608     if (!ad->res_fis || !(pr->cmd & PORT_CMD_FIS_RX)) {
609         return;
610     }
611 
612     /* map cmd_fis */
613     tbl_addr = le64_to_cpu(ad->cur_cmd->tbl_addr);
614     cmd_fis = dma_memory_map(ad->hba->as, tbl_addr, &cmd_len,
615                              DMA_DIRECTION_TO_DEVICE);
616 
617     if (cmd_fis == NULL) {
618         DPRINTF(ad->port_no, "dma_memory_map failed in ahci_write_fis_pio");
619         ahci_trigger_irq(ad->hba, ad, PORT_IRQ_HBUS_ERR);
620         return;
621     }
622 
623     if (cmd_len != 0x80) {
624         DPRINTF(ad->port_no,
625                 "dma_memory_map mapped too few bytes in ahci_write_fis_pio");
626         dma_memory_unmap(ad->hba->as, cmd_fis, cmd_len,
627                          DMA_DIRECTION_TO_DEVICE, cmd_len);
628         ahci_trigger_irq(ad->hba, ad, PORT_IRQ_HBUS_ERR);
629         return;
630     }
631 
632     pio_fis = &ad->res_fis[RES_FIS_PSFIS];
633 
634     pio_fis[0] = SATA_FIS_TYPE_PIO_SETUP;
635     pio_fis[1] = (ad->hba->control_regs.irqstatus ? (1 << 6) : 0);
636     pio_fis[2] = s->status;
637     pio_fis[3] = s->error;
638 
639     pio_fis[4] = s->sector;
640     pio_fis[5] = s->lcyl;
641     pio_fis[6] = s->hcyl;
642     pio_fis[7] = s->select;
643     pio_fis[8] = s->hob_sector;
644     pio_fis[9] = s->hob_lcyl;
645     pio_fis[10] = s->hob_hcyl;
646     pio_fis[11] = 0;
647     pio_fis[12] = cmd_fis[12];
648     pio_fis[13] = cmd_fis[13];
649     pio_fis[14] = 0;
650     pio_fis[15] = s->status;
651     pio_fis[16] = len & 255;
652     pio_fis[17] = len >> 8;
653     pio_fis[18] = 0;
654     pio_fis[19] = 0;
655 
656     /* Update shadow registers: */
657     pr->tfdata = (ad->port.ifs[0].error << 8) |
658         ad->port.ifs[0].status;
659 
660     if (pio_fis[2] & ERR_STAT) {
661         ahci_trigger_irq(ad->hba, ad, PORT_IRQ_TF_ERR);
662     }
663 
664     ahci_trigger_irq(ad->hba, ad, PORT_IRQ_PIOS_FIS);
665 
666     dma_memory_unmap(ad->hba->as, cmd_fis, cmd_len,
667                      DMA_DIRECTION_TO_DEVICE, cmd_len);
668 }
669 
670 static void ahci_write_fis_d2h(AHCIDevice *ad, uint8_t *cmd_fis)
671 {
672     AHCIPortRegs *pr = &ad->port_regs;
673     uint8_t *d2h_fis;
674     int i;
675     dma_addr_t cmd_len = 0x80;
676     int cmd_mapped = 0;
677     IDEState *s = &ad->port.ifs[0];
678 
679     if (!ad->res_fis || !(pr->cmd & PORT_CMD_FIS_RX)) {
680         return;
681     }
682 
683     if (!cmd_fis) {
684         /* map cmd_fis */
685         uint64_t tbl_addr = le64_to_cpu(ad->cur_cmd->tbl_addr);
686         cmd_fis = dma_memory_map(ad->hba->as, tbl_addr, &cmd_len,
687                                  DMA_DIRECTION_TO_DEVICE);
688         cmd_mapped = 1;
689     }
690 
691     d2h_fis = &ad->res_fis[RES_FIS_RFIS];
692 
693     d2h_fis[0] = SATA_FIS_TYPE_REGISTER_D2H;
694     d2h_fis[1] = (ad->hba->control_regs.irqstatus ? (1 << 6) : 0);
695     d2h_fis[2] = s->status;
696     d2h_fis[3] = s->error;
697 
698     d2h_fis[4] = s->sector;
699     d2h_fis[5] = s->lcyl;
700     d2h_fis[6] = s->hcyl;
701     d2h_fis[7] = s->select;
702     d2h_fis[8] = s->hob_sector;
703     d2h_fis[9] = s->hob_lcyl;
704     d2h_fis[10] = s->hob_hcyl;
705     d2h_fis[11] = 0;
706     d2h_fis[12] = cmd_fis[12];
707     d2h_fis[13] = cmd_fis[13];
708     for (i = 14; i < 20; i++) {
709         d2h_fis[i] = 0;
710     }
711 
712     /* Update shadow registers: */
713     pr->tfdata = (ad->port.ifs[0].error << 8) |
714         ad->port.ifs[0].status;
715 
716     if (d2h_fis[2] & ERR_STAT) {
717         ahci_trigger_irq(ad->hba, ad, PORT_IRQ_TF_ERR);
718     }
719 
720     ahci_trigger_irq(ad->hba, ad, PORT_IRQ_D2H_REG_FIS);
721 
722     if (cmd_mapped) {
723         dma_memory_unmap(ad->hba->as, cmd_fis, cmd_len,
724                          DMA_DIRECTION_TO_DEVICE, cmd_len);
725     }
726 }
727 
728 static int prdt_tbl_entry_size(const AHCI_SG *tbl)
729 {
730     return (le32_to_cpu(tbl->flags_size) & AHCI_PRDT_SIZE_MASK) + 1;
731 }
732 
733 static int ahci_populate_sglist(AHCIDevice *ad, QEMUSGList *sglist,
734                                 int32_t offset)
735 {
736     AHCICmdHdr *cmd = ad->cur_cmd;
737     uint32_t opts = le32_to_cpu(cmd->opts);
738     uint64_t prdt_addr = le64_to_cpu(cmd->tbl_addr) + 0x80;
739     int sglist_alloc_hint = opts >> AHCI_CMD_HDR_PRDT_LEN;
740     dma_addr_t prdt_len = (sglist_alloc_hint * sizeof(AHCI_SG));
741     dma_addr_t real_prdt_len = prdt_len;
742     uint8_t *prdt;
743     int i;
744     int r = 0;
745     uint64_t sum = 0;
746     int off_idx = -1;
747     int64_t off_pos = -1;
748     int tbl_entry_size;
749     IDEBus *bus = &ad->port;
750     BusState *qbus = BUS(bus);
751 
752     /*
753      * Note: AHCI PRDT can describe up to 256GiB. SATA/ATA only support
754      * transactions of up to 32MiB as of ATA8-ACS3 rev 1b, assuming a
755      * 512 byte sector size. We limit the PRDT in this implementation to
756      * a reasonably large 2GiB, which can accommodate the maximum transfer
757      * request for sector sizes up to 32K.
758      */
759 
760     if (!sglist_alloc_hint) {
761         DPRINTF(ad->port_no, "no sg list given by guest: 0x%08x\n", opts);
762         return -1;
763     }
764 
765     /* map PRDT */
766     if (!(prdt = dma_memory_map(ad->hba->as, prdt_addr, &prdt_len,
767                                 DMA_DIRECTION_TO_DEVICE))){
768         DPRINTF(ad->port_no, "map failed\n");
769         return -1;
770     }
771 
772     if (prdt_len < real_prdt_len) {
773         DPRINTF(ad->port_no, "mapped less than expected\n");
774         r = -1;
775         goto out;
776     }
777 
778     /* Get entries in the PRDT, init a qemu sglist accordingly */
779     if (sglist_alloc_hint > 0) {
780         AHCI_SG *tbl = (AHCI_SG *)prdt;
781         sum = 0;
782         for (i = 0; i < sglist_alloc_hint; i++) {
783             /* flags_size is zero-based */
784             tbl_entry_size = prdt_tbl_entry_size(&tbl[i]);
785             if (offset <= (sum + tbl_entry_size)) {
786                 off_idx = i;
787                 off_pos = offset - sum;
788                 break;
789             }
790             sum += tbl_entry_size;
791         }
792         if ((off_idx == -1) || (off_pos < 0) || (off_pos > tbl_entry_size)) {
793             DPRINTF(ad->port_no, "%s: Incorrect offset! "
794                             "off_idx: %d, off_pos: %"PRId64"\n",
795                             __func__, off_idx, off_pos);
796             r = -1;
797             goto out;
798         }
799 
800         qemu_sglist_init(sglist, qbus->parent, (sglist_alloc_hint - off_idx),
801                          ad->hba->as);
802         qemu_sglist_add(sglist, le64_to_cpu(tbl[off_idx].addr + off_pos),
803                         prdt_tbl_entry_size(&tbl[off_idx]) - off_pos);
804 
805         for (i = off_idx + 1; i < sglist_alloc_hint; i++) {
806             /* flags_size is zero-based */
807             qemu_sglist_add(sglist, le64_to_cpu(tbl[i].addr),
808                             prdt_tbl_entry_size(&tbl[i]));
809             if (sglist->size > INT32_MAX) {
810                 error_report("AHCI Physical Region Descriptor Table describes "
811                              "more than 2 GiB.\n");
812                 qemu_sglist_destroy(sglist);
813                 r = -1;
814                 goto out;
815             }
816         }
817     }
818 
819 out:
820     dma_memory_unmap(ad->hba->as, prdt, prdt_len,
821                      DMA_DIRECTION_TO_DEVICE, prdt_len);
822     return r;
823 }
824 
825 static void ncq_cb(void *opaque, int ret)
826 {
827     NCQTransferState *ncq_tfs = (NCQTransferState *)opaque;
828     IDEState *ide_state = &ncq_tfs->drive->port.ifs[0];
829 
830     if (ret == -ECANCELED) {
831         return;
832     }
833     /* Clear bit for this tag in SActive */
834     ncq_tfs->drive->port_regs.scr_act &= ~(1 << ncq_tfs->tag);
835 
836     if (ret < 0) {
837         /* error */
838         ide_state->error = ABRT_ERR;
839         ide_state->status = READY_STAT | ERR_STAT;
840         ncq_tfs->drive->port_regs.scr_err |= (1 << ncq_tfs->tag);
841     } else {
842         ide_state->status = READY_STAT | SEEK_STAT;
843     }
844 
845     ahci_write_fis_sdb(ncq_tfs->drive->hba, ncq_tfs->drive->port_no,
846                        (1 << ncq_tfs->tag));
847 
848     DPRINTF(ncq_tfs->drive->port_no, "NCQ transfer tag %d finished\n",
849             ncq_tfs->tag);
850 
851     block_acct_done(blk_get_stats(ncq_tfs->drive->port.ifs[0].blk),
852                     &ncq_tfs->acct);
853     qemu_sglist_destroy(&ncq_tfs->sglist);
854     ncq_tfs->used = 0;
855 }
856 
857 static int is_ncq(uint8_t ata_cmd)
858 {
859     /* Based on SATA 3.2 section 13.6.3.2 */
860     switch (ata_cmd) {
861     case READ_FPDMA_QUEUED:
862     case WRITE_FPDMA_QUEUED:
863     case NCQ_NON_DATA:
864     case RECEIVE_FPDMA_QUEUED:
865     case SEND_FPDMA_QUEUED:
866         return 1;
867     default:
868         return 0;
869     }
870 }
871 
872 static void process_ncq_command(AHCIState *s, int port, uint8_t *cmd_fis,
873                                 int slot)
874 {
875     NCQFrame *ncq_fis = (NCQFrame*)cmd_fis;
876     uint8_t tag = ncq_fis->tag >> 3;
877     NCQTransferState *ncq_tfs = &s->dev[port].ncq_tfs[tag];
878 
879     if (ncq_tfs->used) {
880         /* error - already in use */
881         fprintf(stderr, "%s: tag %d already used\n", __FUNCTION__, tag);
882         return;
883     }
884 
885     ncq_tfs->used = 1;
886     ncq_tfs->drive = &s->dev[port];
887     ncq_tfs->slot = slot;
888     ncq_tfs->lba = ((uint64_t)ncq_fis->lba5 << 40) |
889                    ((uint64_t)ncq_fis->lba4 << 32) |
890                    ((uint64_t)ncq_fis->lba3 << 24) |
891                    ((uint64_t)ncq_fis->lba2 << 16) |
892                    ((uint64_t)ncq_fis->lba1 << 8) |
893                    (uint64_t)ncq_fis->lba0;
894 
895     /* Note: We calculate the sector count, but don't currently rely on it.
896      * The total size of the DMA buffer tells us the transfer size instead. */
897     ncq_tfs->sector_count = ((uint16_t)ncq_fis->sector_count_high << 8) |
898                                 ncq_fis->sector_count_low;
899 
900     DPRINTF(port, "NCQ transfer LBA from %"PRId64" to %"PRId64", "
901             "drive max %"PRId64"\n",
902             ncq_tfs->lba, ncq_tfs->lba + ncq_tfs->sector_count - 2,
903             s->dev[port].port.ifs[0].nb_sectors - 1);
904 
905     ahci_populate_sglist(&s->dev[port], &ncq_tfs->sglist, 0);
906     ncq_tfs->tag = tag;
907 
908     switch(ncq_fis->command) {
909         case READ_FPDMA_QUEUED:
910             DPRINTF(port, "NCQ reading %d sectors from LBA %"PRId64", "
911                     "tag %d\n",
912                     ncq_tfs->sector_count-1, ncq_tfs->lba, ncq_tfs->tag);
913 
914             DPRINTF(port, "tag %d aio read %"PRId64"\n",
915                     ncq_tfs->tag, ncq_tfs->lba);
916 
917             dma_acct_start(ncq_tfs->drive->port.ifs[0].blk, &ncq_tfs->acct,
918                            &ncq_tfs->sglist, BLOCK_ACCT_READ);
919             ncq_tfs->aiocb = dma_blk_read(ncq_tfs->drive->port.ifs[0].blk,
920                                           &ncq_tfs->sglist, ncq_tfs->lba,
921                                           ncq_cb, ncq_tfs);
922             break;
923         case WRITE_FPDMA_QUEUED:
924             DPRINTF(port, "NCQ writing %d sectors to LBA %"PRId64", tag %d\n",
925                     ncq_tfs->sector_count-1, ncq_tfs->lba, ncq_tfs->tag);
926 
927             DPRINTF(port, "tag %d aio write %"PRId64"\n",
928                     ncq_tfs->tag, ncq_tfs->lba);
929 
930             dma_acct_start(ncq_tfs->drive->port.ifs[0].blk, &ncq_tfs->acct,
931                            &ncq_tfs->sglist, BLOCK_ACCT_WRITE);
932             ncq_tfs->aiocb = dma_blk_write(ncq_tfs->drive->port.ifs[0].blk,
933                                            &ncq_tfs->sglist, ncq_tfs->lba,
934                                            ncq_cb, ncq_tfs);
935             break;
936         default:
937             if (is_ncq(cmd_fis[2])) {
938                 DPRINTF(port,
939                         "error: unsupported NCQ command (0x%02x) received\n",
940                         cmd_fis[2]);
941             } else {
942                 DPRINTF(port,
943                         "error: tried to process non-NCQ command as NCQ\n");
944             }
945             qemu_sglist_destroy(&ncq_tfs->sglist);
946     }
947 }
948 
949 static void handle_reg_h2d_fis(AHCIState *s, int port,
950                                int slot, uint8_t *cmd_fis)
951 {
952     IDEState *ide_state = &s->dev[port].port.ifs[0];
953     AHCICmdHdr *cmd = s->dev[port].cur_cmd;
954     uint32_t opts = le32_to_cpu(cmd->opts);
955 
956     if (cmd_fis[1] & 0x0F) {
957         DPRINTF(port, "Port Multiplier not supported."
958                 " cmd_fis[0]=%02x cmd_fis[1]=%02x cmd_fis[2]=%02x\n",
959                 cmd_fis[0], cmd_fis[1], cmd_fis[2]);
960         return;
961     }
962 
963     if (cmd_fis[1] & 0x70) {
964         DPRINTF(port, "Reserved flags set in H2D Register FIS."
965                 " cmd_fis[0]=%02x cmd_fis[1]=%02x cmd_fis[2]=%02x\n",
966                 cmd_fis[0], cmd_fis[1], cmd_fis[2]);
967         return;
968     }
969 
970     if (!(cmd_fis[1] & SATA_FIS_REG_H2D_UPDATE_COMMAND_REGISTER)) {
971         switch (s->dev[port].port_state) {
972         case STATE_RUN:
973             if (cmd_fis[15] & ATA_SRST) {
974                 s->dev[port].port_state = STATE_RESET;
975             }
976             break;
977         case STATE_RESET:
978             if (!(cmd_fis[15] & ATA_SRST)) {
979                 ahci_reset_port(s, port);
980             }
981             break;
982         }
983         return;
984     }
985 
986     /* Check for NCQ command */
987     if (is_ncq(cmd_fis[2])) {
988         process_ncq_command(s, port, cmd_fis, slot);
989         return;
990     }
991 
992     /* Decompose the FIS:
993      * AHCI does not interpret FIS packets, it only forwards them.
994      * SATA 1.0 describes how to decode LBA28 and CHS FIS packets.
995      * Later specifications, e.g, SATA 3.2, describe LBA48 FIS packets.
996      *
997      * ATA4 describes sector number for LBA28/CHS commands.
998      * ATA6 describes sector number for LBA48 commands.
999      * ATA8 deprecates CHS fully, describing only LBA28/48.
1000      *
1001      * We dutifully convert the FIS into IDE registers, and allow the
1002      * core layer to interpret them as needed. */
1003     ide_state->feature = cmd_fis[3];
1004     ide_state->sector = cmd_fis[4];      /* LBA 7:0 */
1005     ide_state->lcyl = cmd_fis[5];        /* LBA 15:8  */
1006     ide_state->hcyl = cmd_fis[6];        /* LBA 23:16 */
1007     ide_state->select = cmd_fis[7];      /* LBA 27:24 (LBA28) */
1008     ide_state->hob_sector = cmd_fis[8];  /* LBA 31:24 */
1009     ide_state->hob_lcyl = cmd_fis[9];    /* LBA 39:32 */
1010     ide_state->hob_hcyl = cmd_fis[10];   /* LBA 47:40 */
1011     ide_state->hob_feature = cmd_fis[11];
1012     ide_state->nsector = (int64_t)((cmd_fis[13] << 8) | cmd_fis[12]);
1013     /* 14, 16, 17, 18, 19: Reserved (SATA 1.0) */
1014     /* 15: Only valid when UPDATE_COMMAND not set. */
1015 
1016     /* Copy the ACMD field (ATAPI packet, if any) from the AHCI command
1017      * table to ide_state->io_buffer */
1018     if (opts & AHCI_CMD_ATAPI) {
1019         memcpy(ide_state->io_buffer, &cmd_fis[AHCI_COMMAND_TABLE_ACMD], 0x10);
1020         debug_print_fis(ide_state->io_buffer, 0x10);
1021         s->dev[port].done_atapi_packet = false;
1022         /* XXX send PIO setup FIS */
1023     }
1024 
1025     ide_state->error = 0;
1026 
1027     /* Reset transferred byte counter */
1028     cmd->status = 0;
1029 
1030     /* We're ready to process the command in FIS byte 2. */
1031     ide_exec_cmd(&s->dev[port].port, cmd_fis[2]);
1032 }
1033 
1034 static int handle_cmd(AHCIState *s, int port, int slot)
1035 {
1036     IDEState *ide_state;
1037     uint64_t tbl_addr;
1038     AHCICmdHdr *cmd;
1039     uint8_t *cmd_fis;
1040     dma_addr_t cmd_len;
1041 
1042     if (s->dev[port].port.ifs[0].status & (BUSY_STAT|DRQ_STAT)) {
1043         /* Engine currently busy, try again later */
1044         DPRINTF(port, "engine busy\n");
1045         return -1;
1046     }
1047 
1048     if (!s->dev[port].lst) {
1049         DPRINTF(port, "error: lst not given but cmd handled");
1050         return -1;
1051     }
1052     cmd = &((AHCICmdHdr *)s->dev[port].lst)[slot];
1053     /* remember current slot handle for later */
1054     s->dev[port].cur_cmd = cmd;
1055 
1056     /* The device we are working for */
1057     ide_state = &s->dev[port].port.ifs[0];
1058     if (!ide_state->blk) {
1059         DPRINTF(port, "error: guest accessed unused port");
1060         return -1;
1061     }
1062 
1063     tbl_addr = le64_to_cpu(cmd->tbl_addr);
1064     cmd_len = 0x80;
1065     cmd_fis = dma_memory_map(s->as, tbl_addr, &cmd_len,
1066                              DMA_DIRECTION_FROM_DEVICE);
1067     if (!cmd_fis) {
1068         DPRINTF(port, "error: guest passed us an invalid cmd fis\n");
1069         return -1;
1070     } else if (cmd_len != 0x80) {
1071         ahci_trigger_irq(s, &s->dev[port], PORT_IRQ_HBUS_ERR);
1072         DPRINTF(port, "error: dma_memory_map failed: "
1073                 "(len(%02"PRIx64") != 0x80)\n",
1074                 cmd_len);
1075         goto out;
1076     }
1077     debug_print_fis(cmd_fis, 0x80);
1078 
1079     switch (cmd_fis[0]) {
1080         case SATA_FIS_TYPE_REGISTER_H2D:
1081             handle_reg_h2d_fis(s, port, slot, cmd_fis);
1082             break;
1083         default:
1084             DPRINTF(port, "unknown command cmd_fis[0]=%02x cmd_fis[1]=%02x "
1085                           "cmd_fis[2]=%02x\n", cmd_fis[0], cmd_fis[1],
1086                           cmd_fis[2]);
1087             break;
1088     }
1089 
1090 out:
1091     dma_memory_unmap(s->as, cmd_fis, cmd_len, DMA_DIRECTION_FROM_DEVICE,
1092                      cmd_len);
1093 
1094     if (s->dev[port].port.ifs[0].status & (BUSY_STAT|DRQ_STAT)) {
1095         /* async command, complete later */
1096         s->dev[port].busy_slot = slot;
1097         return -1;
1098     }
1099 
1100     /* done handling the command */
1101     return 0;
1102 }
1103 
1104 /* DMA dev <-> ram */
1105 static void ahci_start_transfer(IDEDMA *dma)
1106 {
1107     AHCIDevice *ad = DO_UPCAST(AHCIDevice, dma, dma);
1108     IDEState *s = &ad->port.ifs[0];
1109     uint32_t size = (uint32_t)(s->data_end - s->data_ptr);
1110     /* write == ram -> device */
1111     uint32_t opts = le32_to_cpu(ad->cur_cmd->opts);
1112     int is_write = opts & AHCI_CMD_WRITE;
1113     int is_atapi = opts & AHCI_CMD_ATAPI;
1114     int has_sglist = 0;
1115 
1116     if (is_atapi && !ad->done_atapi_packet) {
1117         /* already prepopulated iobuffer */
1118         ad->done_atapi_packet = true;
1119         size = 0;
1120         goto out;
1121     }
1122 
1123     if (ahci_dma_prepare_buf(dma, is_write)) {
1124         has_sglist = 1;
1125     }
1126 
1127     DPRINTF(ad->port_no, "%sing %d bytes on %s w/%s sglist\n",
1128             is_write ? "writ" : "read", size, is_atapi ? "atapi" : "ata",
1129             has_sglist ? "" : "o");
1130 
1131     if (has_sglist && size) {
1132         if (is_write) {
1133             dma_buf_write(s->data_ptr, size, &s->sg);
1134         } else {
1135             dma_buf_read(s->data_ptr, size, &s->sg);
1136         }
1137     }
1138 
1139 out:
1140     /* declare that we processed everything */
1141     s->data_ptr = s->data_end;
1142 
1143     /* Update number of transferred bytes, destroy sglist */
1144     ahci_commit_buf(dma, size);
1145 
1146     s->end_transfer_func(s);
1147 
1148     if (!(s->status & DRQ_STAT)) {
1149         /* done with PIO send/receive */
1150         ahci_write_fis_pio(ad, le32_to_cpu(ad->cur_cmd->status));
1151     }
1152 }
1153 
1154 static void ahci_start_dma(IDEDMA *dma, IDEState *s,
1155                            BlockCompletionFunc *dma_cb)
1156 {
1157     AHCIDevice *ad = DO_UPCAST(AHCIDevice, dma, dma);
1158     DPRINTF(ad->port_no, "\n");
1159     s->io_buffer_offset = 0;
1160     dma_cb(s, 0);
1161 }
1162 
1163 /**
1164  * Called in DMA R/W chains to read the PRDT, utilizing ahci_populate_sglist.
1165  * Not currently invoked by PIO R/W chains,
1166  * which invoke ahci_populate_sglist via ahci_start_transfer.
1167  */
1168 static int32_t ahci_dma_prepare_buf(IDEDMA *dma, int is_write)
1169 {
1170     AHCIDevice *ad = DO_UPCAST(AHCIDevice, dma, dma);
1171     IDEState *s = &ad->port.ifs[0];
1172 
1173     if (ahci_populate_sglist(ad, &s->sg, s->io_buffer_offset) == -1) {
1174         DPRINTF(ad->port_no, "ahci_dma_prepare_buf failed.\n");
1175         return -1;
1176     }
1177     s->io_buffer_size = s->sg.size;
1178 
1179     DPRINTF(ad->port_no, "len=%#x\n", s->io_buffer_size);
1180     return s->io_buffer_size;
1181 }
1182 
1183 /**
1184  * Destroys the scatter-gather list,
1185  * and updates the command header with a bytes-read value.
1186  * called explicitly via ahci_dma_rw_buf (ATAPI DMA),
1187  * and ahci_start_transfer (PIO R/W),
1188  * and called via callback from ide_dma_cb for DMA R/W paths.
1189  */
1190 static void ahci_commit_buf(IDEDMA *dma, uint32_t tx_bytes)
1191 {
1192     AHCIDevice *ad = DO_UPCAST(AHCIDevice, dma, dma);
1193     IDEState *s = &ad->port.ifs[0];
1194 
1195     tx_bytes += le32_to_cpu(ad->cur_cmd->status);
1196     ad->cur_cmd->status = cpu_to_le32(tx_bytes);
1197 
1198     qemu_sglist_destroy(&s->sg);
1199 }
1200 
1201 static int ahci_dma_rw_buf(IDEDMA *dma, int is_write)
1202 {
1203     AHCIDevice *ad = DO_UPCAST(AHCIDevice, dma, dma);
1204     IDEState *s = &ad->port.ifs[0];
1205     uint8_t *p = s->io_buffer + s->io_buffer_index;
1206     int l = s->io_buffer_size - s->io_buffer_index;
1207 
1208     if (ahci_populate_sglist(ad, &s->sg, s->io_buffer_offset)) {
1209         return 0;
1210     }
1211 
1212     if (is_write) {
1213         dma_buf_read(p, l, &s->sg);
1214     } else {
1215         dma_buf_write(p, l, &s->sg);
1216     }
1217 
1218     /* free sglist, update byte count */
1219     ahci_commit_buf(dma, l);
1220 
1221     s->io_buffer_index += l;
1222     s->io_buffer_offset += l;
1223 
1224     DPRINTF(ad->port_no, "len=%#x\n", l);
1225 
1226     return 1;
1227 }
1228 
1229 static int ahci_dma_set_unit(IDEDMA *dma, int unit)
1230 {
1231     /* only a single unit per link */
1232     return 0;
1233 }
1234 
1235 static void ahci_cmd_done(IDEDMA *dma)
1236 {
1237     AHCIDevice *ad = DO_UPCAST(AHCIDevice, dma, dma);
1238 
1239     DPRINTF(ad->port_no, "cmd done\n");
1240 
1241     /* update d2h status */
1242     ahci_write_fis_d2h(ad, NULL);
1243 
1244     if (!ad->check_bh) {
1245         /* maybe we still have something to process, check later */
1246         ad->check_bh = qemu_bh_new(ahci_check_cmd_bh, ad);
1247         qemu_bh_schedule(ad->check_bh);
1248     }
1249 }
1250 
1251 static void ahci_irq_set(void *opaque, int n, int level)
1252 {
1253 }
1254 
1255 static void ahci_dma_restart_cb(void *opaque, int running, RunState state)
1256 {
1257 }
1258 
1259 static const IDEDMAOps ahci_dma_ops = {
1260     .start_dma = ahci_start_dma,
1261     .start_transfer = ahci_start_transfer,
1262     .prepare_buf = ahci_dma_prepare_buf,
1263     .commit_buf = ahci_commit_buf,
1264     .rw_buf = ahci_dma_rw_buf,
1265     .set_unit = ahci_dma_set_unit,
1266     .cmd_done = ahci_cmd_done,
1267     .restart_cb = ahci_dma_restart_cb,
1268 };
1269 
1270 void ahci_init(AHCIState *s, DeviceState *qdev, AddressSpace *as, int ports)
1271 {
1272     qemu_irq *irqs;
1273     int i;
1274 
1275     s->as = as;
1276     s->ports = ports;
1277     s->dev = g_new0(AHCIDevice, ports);
1278     ahci_reg_init(s);
1279     /* XXX BAR size should be 1k, but that breaks, so bump it to 4k for now */
1280     memory_region_init_io(&s->mem, OBJECT(qdev), &ahci_mem_ops, s,
1281                           "ahci", AHCI_MEM_BAR_SIZE);
1282     memory_region_init_io(&s->idp, OBJECT(qdev), &ahci_idp_ops, s,
1283                           "ahci-idp", 32);
1284 
1285     irqs = qemu_allocate_irqs(ahci_irq_set, s, s->ports);
1286 
1287     for (i = 0; i < s->ports; i++) {
1288         AHCIDevice *ad = &s->dev[i];
1289 
1290         ide_bus_new(&ad->port, sizeof(ad->port), qdev, i, 1);
1291         ide_init2(&ad->port, irqs[i]);
1292 
1293         ad->hba = s;
1294         ad->port_no = i;
1295         ad->port.dma = &ad->dma;
1296         ad->port.dma->ops = &ahci_dma_ops;
1297     }
1298 }
1299 
1300 void ahci_uninit(AHCIState *s)
1301 {
1302     g_free(s->dev);
1303 }
1304 
1305 void ahci_reset(AHCIState *s)
1306 {
1307     AHCIPortRegs *pr;
1308     int i;
1309 
1310     s->control_regs.irqstatus = 0;
1311     /* AHCI Enable (AE)
1312      * The implementation of this bit is dependent upon the value of the
1313      * CAP.SAM bit. If CAP.SAM is '0', then GHC.AE shall be read-write and
1314      * shall have a reset value of '0'. If CAP.SAM is '1', then AE shall be
1315      * read-only and shall have a reset value of '1'.
1316      *
1317      * We set HOST_CAP_AHCI so we must enable AHCI at reset.
1318      */
1319     s->control_regs.ghc = HOST_CTL_AHCI_EN;
1320 
1321     for (i = 0; i < s->ports; i++) {
1322         pr = &s->dev[i].port_regs;
1323         pr->irq_stat = 0;
1324         pr->irq_mask = 0;
1325         pr->scr_ctl = 0;
1326         pr->cmd = PORT_CMD_SPIN_UP | PORT_CMD_POWER_ON;
1327         ahci_reset_port(s, i);
1328     }
1329 }
1330 
1331 static const VMStateDescription vmstate_ahci_device = {
1332     .name = "ahci port",
1333     .version_id = 1,
1334     .fields = (VMStateField[]) {
1335         VMSTATE_IDE_BUS(port, AHCIDevice),
1336         VMSTATE_UINT32(port_state, AHCIDevice),
1337         VMSTATE_UINT32(finished, AHCIDevice),
1338         VMSTATE_UINT32(port_regs.lst_addr, AHCIDevice),
1339         VMSTATE_UINT32(port_regs.lst_addr_hi, AHCIDevice),
1340         VMSTATE_UINT32(port_regs.fis_addr, AHCIDevice),
1341         VMSTATE_UINT32(port_regs.fis_addr_hi, AHCIDevice),
1342         VMSTATE_UINT32(port_regs.irq_stat, AHCIDevice),
1343         VMSTATE_UINT32(port_regs.irq_mask, AHCIDevice),
1344         VMSTATE_UINT32(port_regs.cmd, AHCIDevice),
1345         VMSTATE_UINT32(port_regs.tfdata, AHCIDevice),
1346         VMSTATE_UINT32(port_regs.sig, AHCIDevice),
1347         VMSTATE_UINT32(port_regs.scr_stat, AHCIDevice),
1348         VMSTATE_UINT32(port_regs.scr_ctl, AHCIDevice),
1349         VMSTATE_UINT32(port_regs.scr_err, AHCIDevice),
1350         VMSTATE_UINT32(port_regs.scr_act, AHCIDevice),
1351         VMSTATE_UINT32(port_regs.cmd_issue, AHCIDevice),
1352         VMSTATE_BOOL(done_atapi_packet, AHCIDevice),
1353         VMSTATE_INT32(busy_slot, AHCIDevice),
1354         VMSTATE_BOOL(init_d2h_sent, AHCIDevice),
1355         VMSTATE_END_OF_LIST()
1356     },
1357 };
1358 
1359 static int ahci_state_post_load(void *opaque, int version_id)
1360 {
1361     int i;
1362     struct AHCIDevice *ad;
1363     AHCIState *s = opaque;
1364 
1365     for (i = 0; i < s->ports; i++) {
1366         ad = &s->dev[i];
1367         AHCIPortRegs *pr = &ad->port_regs;
1368 
1369         map_page(s->as, &ad->lst,
1370                  ((uint64_t)pr->lst_addr_hi << 32) | pr->lst_addr, 1024);
1371         map_page(s->as, &ad->res_fis,
1372                  ((uint64_t)pr->fis_addr_hi << 32) | pr->fis_addr, 256);
1373         /*
1374          * All pending i/o should be flushed out on a migrate. However,
1375          * we might not have cleared the busy_slot since this is done
1376          * in a bh. Also, issue i/o against any slots that are pending.
1377          */
1378         if ((ad->busy_slot != -1) &&
1379             !(ad->port.ifs[0].status & (BUSY_STAT|DRQ_STAT))) {
1380             pr->cmd_issue &= ~(1 << ad->busy_slot);
1381             ad->busy_slot = -1;
1382         }
1383         check_cmd(s, i);
1384     }
1385 
1386     return 0;
1387 }
1388 
1389 const VMStateDescription vmstate_ahci = {
1390     .name = "ahci",
1391     .version_id = 1,
1392     .post_load = ahci_state_post_load,
1393     .fields = (VMStateField[]) {
1394         VMSTATE_STRUCT_VARRAY_POINTER_INT32(dev, AHCIState, ports,
1395                                      vmstate_ahci_device, AHCIDevice),
1396         VMSTATE_UINT32(control_regs.cap, AHCIState),
1397         VMSTATE_UINT32(control_regs.ghc, AHCIState),
1398         VMSTATE_UINT32(control_regs.irqstatus, AHCIState),
1399         VMSTATE_UINT32(control_regs.impl, AHCIState),
1400         VMSTATE_UINT32(control_regs.version, AHCIState),
1401         VMSTATE_UINT32(idp_index, AHCIState),
1402         VMSTATE_INT32_EQUAL(ports, AHCIState),
1403         VMSTATE_END_OF_LIST()
1404     },
1405 };
1406 
1407 #define TYPE_SYSBUS_AHCI "sysbus-ahci"
1408 #define SYSBUS_AHCI(obj) OBJECT_CHECK(SysbusAHCIState, (obj), TYPE_SYSBUS_AHCI)
1409 
1410 typedef struct SysbusAHCIState {
1411     /*< private >*/
1412     SysBusDevice parent_obj;
1413     /*< public >*/
1414 
1415     AHCIState ahci;
1416     uint32_t num_ports;
1417 } SysbusAHCIState;
1418 
1419 static const VMStateDescription vmstate_sysbus_ahci = {
1420     .name = "sysbus-ahci",
1421     .unmigratable = 1, /* Still buggy under I/O load */
1422     .fields = (VMStateField[]) {
1423         VMSTATE_AHCI(ahci, SysbusAHCIState),
1424         VMSTATE_END_OF_LIST()
1425     },
1426 };
1427 
1428 static void sysbus_ahci_reset(DeviceState *dev)
1429 {
1430     SysbusAHCIState *s = SYSBUS_AHCI(dev);
1431 
1432     ahci_reset(&s->ahci);
1433 }
1434 
1435 static void sysbus_ahci_realize(DeviceState *dev, Error **errp)
1436 {
1437     SysBusDevice *sbd = SYS_BUS_DEVICE(dev);
1438     SysbusAHCIState *s = SYSBUS_AHCI(dev);
1439 
1440     ahci_init(&s->ahci, dev, &address_space_memory, s->num_ports);
1441 
1442     sysbus_init_mmio(sbd, &s->ahci.mem);
1443     sysbus_init_irq(sbd, &s->ahci.irq);
1444 }
1445 
1446 static Property sysbus_ahci_properties[] = {
1447     DEFINE_PROP_UINT32("num-ports", SysbusAHCIState, num_ports, 1),
1448     DEFINE_PROP_END_OF_LIST(),
1449 };
1450 
1451 static void sysbus_ahci_class_init(ObjectClass *klass, void *data)
1452 {
1453     DeviceClass *dc = DEVICE_CLASS(klass);
1454 
1455     dc->realize = sysbus_ahci_realize;
1456     dc->vmsd = &vmstate_sysbus_ahci;
1457     dc->props = sysbus_ahci_properties;
1458     dc->reset = sysbus_ahci_reset;
1459     set_bit(DEVICE_CATEGORY_STORAGE, dc->categories);
1460 }
1461 
1462 static const TypeInfo sysbus_ahci_info = {
1463     .name          = TYPE_SYSBUS_AHCI,
1464     .parent        = TYPE_SYS_BUS_DEVICE,
1465     .instance_size = sizeof(SysbusAHCIState),
1466     .class_init    = sysbus_ahci_class_init,
1467 };
1468 
1469 static void sysbus_ahci_register_types(void)
1470 {
1471     type_register_static(&sysbus_ahci_info);
1472 }
1473 
1474 type_init(sysbus_ahci_register_types)
1475 
1476 void ahci_ide_create_devs(PCIDevice *dev, DriveInfo **hd)
1477 {
1478     AHCIPCIState *d = ICH_AHCI(dev);
1479     AHCIState *ahci = &d->ahci;
1480     int i;
1481 
1482     for (i = 0; i < ahci->ports; i++) {
1483         if (hd[i] == NULL) {
1484             continue;
1485         }
1486         ide_create_drive(&ahci->dev[i].port, 0, hd[i]);
1487     }
1488 
1489 }
1490