11fd7a697STejun Heo /* 21fd7a697STejun Heo * sata_inic162x.c - Driver for Initio 162x SATA controllers 31fd7a697STejun Heo * 41fd7a697STejun Heo * Copyright 2006 SUSE Linux Products GmbH 51fd7a697STejun Heo * Copyright 2006 Tejun Heo <teheo@novell.com> 61fd7a697STejun Heo * 71fd7a697STejun Heo * This file is released under GPL v2. 81fd7a697STejun Heo * 91fd7a697STejun Heo * This controller is eccentric and easily locks up if something isn't 101fd7a697STejun Heo * right. Documentation is available at initio's website but it only 111fd7a697STejun Heo * documents registers (not programming model). 121fd7a697STejun Heo * 131fd7a697STejun Heo * - ATA disks work. 141fd7a697STejun Heo * - Hotplug works. 151fd7a697STejun Heo * - ATAPI read works but burning doesn't. This thing is really 161fd7a697STejun Heo * peculiar about ATAPI and I couldn't figure out how ATAPI PIO and 171fd7a697STejun Heo * ATAPI DMA WRITE should be programmed. If you've got a clue, be 181fd7a697STejun Heo * my guest. 191fd7a697STejun Heo * - Both STR and STD work. 201fd7a697STejun Heo */ 211fd7a697STejun Heo 221fd7a697STejun Heo #include <linux/kernel.h> 231fd7a697STejun Heo #include <linux/module.h> 241fd7a697STejun Heo #include <linux/pci.h> 251fd7a697STejun Heo #include <scsi/scsi_host.h> 261fd7a697STejun Heo #include <linux/libata.h> 271fd7a697STejun Heo #include <linux/blkdev.h> 281fd7a697STejun Heo #include <scsi/scsi_device.h> 291fd7a697STejun Heo 301fd7a697STejun Heo #define DRV_NAME "sata_inic162x" 312a3103ceSJeff Garzik #define DRV_VERSION "0.3" 321fd7a697STejun Heo 331fd7a697STejun Heo enum { 341fd7a697STejun Heo MMIO_BAR = 5, 351fd7a697STejun Heo 361fd7a697STejun Heo NR_PORTS = 2, 371fd7a697STejun Heo 381fd7a697STejun Heo HOST_CTL = 0x7c, 391fd7a697STejun Heo HOST_STAT = 0x7e, 401fd7a697STejun Heo HOST_IRQ_STAT = 0xbc, 411fd7a697STejun Heo HOST_IRQ_MASK = 0xbe, 421fd7a697STejun Heo 431fd7a697STejun Heo PORT_SIZE = 0x40, 441fd7a697STejun Heo 451fd7a697STejun Heo /* registers for ATA TF operation */ 461fd7a697STejun Heo PORT_TF = 0x00, 471fd7a697STejun Heo PORT_ALT_STAT = 0x08, 481fd7a697STejun Heo PORT_IRQ_STAT = 0x09, 491fd7a697STejun Heo PORT_IRQ_MASK = 0x0a, 501fd7a697STejun Heo PORT_PRD_CTL = 0x0b, 511fd7a697STejun Heo PORT_PRD_ADDR = 0x0c, 521fd7a697STejun Heo PORT_PRD_XFERLEN = 0x10, 531fd7a697STejun Heo 541fd7a697STejun Heo /* IDMA register */ 551fd7a697STejun Heo PORT_IDMA_CTL = 0x14, 561fd7a697STejun Heo 571fd7a697STejun Heo PORT_SCR = 0x20, 581fd7a697STejun Heo 591fd7a697STejun Heo /* HOST_CTL bits */ 601fd7a697STejun Heo HCTL_IRQOFF = (1 << 8), /* global IRQ off */ 611fd7a697STejun Heo HCTL_PWRDWN = (1 << 13), /* power down PHYs */ 621fd7a697STejun Heo HCTL_SOFTRST = (1 << 13), /* global reset (no phy reset) */ 631fd7a697STejun Heo HCTL_RPGSEL = (1 << 15), /* register page select */ 641fd7a697STejun Heo 651fd7a697STejun Heo HCTL_KNOWN_BITS = HCTL_IRQOFF | HCTL_PWRDWN | HCTL_SOFTRST | 661fd7a697STejun Heo HCTL_RPGSEL, 671fd7a697STejun Heo 681fd7a697STejun Heo /* HOST_IRQ_(STAT|MASK) bits */ 691fd7a697STejun Heo HIRQ_PORT0 = (1 << 0), 701fd7a697STejun Heo HIRQ_PORT1 = (1 << 1), 711fd7a697STejun Heo HIRQ_SOFT = (1 << 14), 721fd7a697STejun Heo HIRQ_GLOBAL = (1 << 15), /* STAT only */ 731fd7a697STejun Heo 741fd7a697STejun Heo /* PORT_IRQ_(STAT|MASK) bits */ 751fd7a697STejun Heo PIRQ_OFFLINE = (1 << 0), /* device unplugged */ 761fd7a697STejun Heo PIRQ_ONLINE = (1 << 1), /* device plugged */ 771fd7a697STejun Heo PIRQ_COMPLETE = (1 << 2), /* completion interrupt */ 781fd7a697STejun Heo PIRQ_FATAL = (1 << 3), /* fatal error */ 791fd7a697STejun Heo PIRQ_ATA = (1 << 4), /* ATA interrupt */ 801fd7a697STejun Heo PIRQ_REPLY = (1 << 5), /* reply FIFO not empty */ 811fd7a697STejun Heo PIRQ_PENDING = (1 << 7), /* port IRQ pending (STAT only) */ 821fd7a697STejun Heo 831fd7a697STejun Heo PIRQ_ERR = PIRQ_OFFLINE | PIRQ_ONLINE | PIRQ_FATAL, 841fd7a697STejun Heo 851fd7a697STejun Heo PIRQ_MASK_DMA_READ = PIRQ_REPLY | PIRQ_ATA, 861fd7a697STejun Heo PIRQ_MASK_OTHER = PIRQ_REPLY | PIRQ_COMPLETE, 871fd7a697STejun Heo PIRQ_MASK_FREEZE = 0xff, 881fd7a697STejun Heo 891fd7a697STejun Heo /* PORT_PRD_CTL bits */ 901fd7a697STejun Heo PRD_CTL_START = (1 << 0), 911fd7a697STejun Heo PRD_CTL_WR = (1 << 3), 921fd7a697STejun Heo PRD_CTL_DMAEN = (1 << 7), /* DMA enable */ 931fd7a697STejun Heo 941fd7a697STejun Heo /* PORT_IDMA_CTL bits */ 951fd7a697STejun Heo IDMA_CTL_RST_ATA = (1 << 2), /* hardreset ATA bus */ 961fd7a697STejun Heo IDMA_CTL_RST_IDMA = (1 << 5), /* reset IDMA machinary */ 971fd7a697STejun Heo IDMA_CTL_GO = (1 << 7), /* IDMA mode go */ 981fd7a697STejun Heo IDMA_CTL_ATA_NIEN = (1 << 8), /* ATA IRQ disable */ 991fd7a697STejun Heo }; 1001fd7a697STejun Heo 1011fd7a697STejun Heo struct inic_host_priv { 1021fd7a697STejun Heo u16 cached_hctl; 1031fd7a697STejun Heo }; 1041fd7a697STejun Heo 1051fd7a697STejun Heo struct inic_port_priv { 1061fd7a697STejun Heo u8 dfl_prdctl; 1071fd7a697STejun Heo u8 cached_prdctl; 1081fd7a697STejun Heo u8 cached_pirq_mask; 1091fd7a697STejun Heo }; 1101fd7a697STejun Heo 1111fd7a697STejun Heo static struct scsi_host_template inic_sht = { 11268d1d07bSTejun Heo ATA_BMDMA_SHT(DRV_NAME), 1131fd7a697STejun Heo }; 1141fd7a697STejun Heo 1151fd7a697STejun Heo static const int scr_map[] = { 1161fd7a697STejun Heo [SCR_STATUS] = 0, 1171fd7a697STejun Heo [SCR_ERROR] = 1, 1181fd7a697STejun Heo [SCR_CONTROL] = 2, 1191fd7a697STejun Heo }; 1201fd7a697STejun Heo 1211fd7a697STejun Heo static void __iomem *inic_port_base(struct ata_port *ap) 1221fd7a697STejun Heo { 1230d5ff566STejun Heo return ap->host->iomap[MMIO_BAR] + ap->port_no * PORT_SIZE; 1241fd7a697STejun Heo } 1251fd7a697STejun Heo 1261fd7a697STejun Heo static void __inic_set_pirq_mask(struct ata_port *ap, u8 mask) 1271fd7a697STejun Heo { 1281fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 1291fd7a697STejun Heo struct inic_port_priv *pp = ap->private_data; 1301fd7a697STejun Heo 1311fd7a697STejun Heo writeb(mask, port_base + PORT_IRQ_MASK); 1321fd7a697STejun Heo pp->cached_pirq_mask = mask; 1331fd7a697STejun Heo } 1341fd7a697STejun Heo 1351fd7a697STejun Heo static void inic_set_pirq_mask(struct ata_port *ap, u8 mask) 1361fd7a697STejun Heo { 1371fd7a697STejun Heo struct inic_port_priv *pp = ap->private_data; 1381fd7a697STejun Heo 1391fd7a697STejun Heo if (pp->cached_pirq_mask != mask) 1401fd7a697STejun Heo __inic_set_pirq_mask(ap, mask); 1411fd7a697STejun Heo } 1421fd7a697STejun Heo 1431fd7a697STejun Heo static void inic_reset_port(void __iomem *port_base) 1441fd7a697STejun Heo { 1451fd7a697STejun Heo void __iomem *idma_ctl = port_base + PORT_IDMA_CTL; 1461fd7a697STejun Heo u16 ctl; 1471fd7a697STejun Heo 1481fd7a697STejun Heo ctl = readw(idma_ctl); 1491fd7a697STejun Heo ctl &= ~(IDMA_CTL_RST_IDMA | IDMA_CTL_ATA_NIEN | IDMA_CTL_GO); 1501fd7a697STejun Heo 1511fd7a697STejun Heo /* mask IRQ and assert reset */ 1521fd7a697STejun Heo writew(ctl | IDMA_CTL_RST_IDMA | IDMA_CTL_ATA_NIEN, idma_ctl); 1531fd7a697STejun Heo readw(idma_ctl); /* flush */ 1541fd7a697STejun Heo 1551fd7a697STejun Heo /* give it some time */ 1561fd7a697STejun Heo msleep(1); 1571fd7a697STejun Heo 1581fd7a697STejun Heo /* release reset */ 1591fd7a697STejun Heo writew(ctl | IDMA_CTL_ATA_NIEN, idma_ctl); 1601fd7a697STejun Heo 1611fd7a697STejun Heo /* clear irq */ 1621fd7a697STejun Heo writeb(0xff, port_base + PORT_IRQ_STAT); 1631fd7a697STejun Heo 1641fd7a697STejun Heo /* reenable ATA IRQ, turn off IDMA mode */ 1651fd7a697STejun Heo writew(ctl, idma_ctl); 1661fd7a697STejun Heo } 1671fd7a697STejun Heo 168da3dbb17STejun Heo static int inic_scr_read(struct ata_port *ap, unsigned sc_reg, u32 *val) 1691fd7a697STejun Heo { 17059f99880SJeff Garzik void __iomem *scr_addr = ap->ioaddr.scr_addr; 1711fd7a697STejun Heo void __iomem *addr; 1721fd7a697STejun Heo 1731fd7a697STejun Heo if (unlikely(sc_reg >= ARRAY_SIZE(scr_map))) 174da3dbb17STejun Heo return -EINVAL; 1751fd7a697STejun Heo 1761fd7a697STejun Heo addr = scr_addr + scr_map[sc_reg] * 4; 177da3dbb17STejun Heo *val = readl(scr_addr + scr_map[sc_reg] * 4); 1781fd7a697STejun Heo 1791fd7a697STejun Heo /* this controller has stuck DIAG.N, ignore it */ 1801fd7a697STejun Heo if (sc_reg == SCR_ERROR) 181da3dbb17STejun Heo *val &= ~SERR_PHYRDY_CHG; 182da3dbb17STejun Heo return 0; 1831fd7a697STejun Heo } 1841fd7a697STejun Heo 185da3dbb17STejun Heo static int inic_scr_write(struct ata_port *ap, unsigned sc_reg, u32 val) 1861fd7a697STejun Heo { 18759f99880SJeff Garzik void __iomem *scr_addr = ap->ioaddr.scr_addr; 1881fd7a697STejun Heo void __iomem *addr; 1891fd7a697STejun Heo 1901fd7a697STejun Heo if (unlikely(sc_reg >= ARRAY_SIZE(scr_map))) 191da3dbb17STejun Heo return -EINVAL; 1921fd7a697STejun Heo 1931fd7a697STejun Heo addr = scr_addr + scr_map[sc_reg] * 4; 1941fd7a697STejun Heo writel(val, scr_addr + scr_map[sc_reg] * 4); 195da3dbb17STejun Heo return 0; 1961fd7a697STejun Heo } 1971fd7a697STejun Heo 1981fd7a697STejun Heo /* 1991fd7a697STejun Heo * In TF mode, inic162x is very similar to SFF device. TF registers 2001fd7a697STejun Heo * function the same. DMA engine behaves similary using the same PRD 2011fd7a697STejun Heo * format as BMDMA but different command register, interrupt and event 2021fd7a697STejun Heo * notification methods are used. The following inic_bmdma_*() 2031fd7a697STejun Heo * functions do the impedance matching. 2041fd7a697STejun Heo */ 2051fd7a697STejun Heo static void inic_bmdma_setup(struct ata_queued_cmd *qc) 2061fd7a697STejun Heo { 2071fd7a697STejun Heo struct ata_port *ap = qc->ap; 2081fd7a697STejun Heo struct inic_port_priv *pp = ap->private_data; 2091fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 2101fd7a697STejun Heo int rw = qc->tf.flags & ATA_TFLAG_WRITE; 2111fd7a697STejun Heo 2121fd7a697STejun Heo /* make sure device sees PRD table writes */ 2131fd7a697STejun Heo wmb(); 2141fd7a697STejun Heo 2151fd7a697STejun Heo /* load transfer length */ 2161fd7a697STejun Heo writel(qc->nbytes, port_base + PORT_PRD_XFERLEN); 2171fd7a697STejun Heo 2181fd7a697STejun Heo /* turn on DMA and specify data direction */ 2191fd7a697STejun Heo pp->cached_prdctl = pp->dfl_prdctl | PRD_CTL_DMAEN; 2201fd7a697STejun Heo if (!rw) 2211fd7a697STejun Heo pp->cached_prdctl |= PRD_CTL_WR; 2221fd7a697STejun Heo writeb(pp->cached_prdctl, port_base + PORT_PRD_CTL); 2231fd7a697STejun Heo 2241fd7a697STejun Heo /* issue r/w command */ 2251fd7a697STejun Heo ap->ops->exec_command(ap, &qc->tf); 2261fd7a697STejun Heo } 2271fd7a697STejun Heo 2281fd7a697STejun Heo static void inic_bmdma_start(struct ata_queued_cmd *qc) 2291fd7a697STejun Heo { 2301fd7a697STejun Heo struct ata_port *ap = qc->ap; 2311fd7a697STejun Heo struct inic_port_priv *pp = ap->private_data; 2321fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 2331fd7a697STejun Heo 2341fd7a697STejun Heo /* start host DMA transaction */ 2351fd7a697STejun Heo pp->cached_prdctl |= PRD_CTL_START; 2361fd7a697STejun Heo writeb(pp->cached_prdctl, port_base + PORT_PRD_CTL); 2371fd7a697STejun Heo } 2381fd7a697STejun Heo 2391fd7a697STejun Heo static void inic_bmdma_stop(struct ata_queued_cmd *qc) 2401fd7a697STejun Heo { 2411fd7a697STejun Heo struct ata_port *ap = qc->ap; 2421fd7a697STejun Heo struct inic_port_priv *pp = ap->private_data; 2431fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 2441fd7a697STejun Heo 2451fd7a697STejun Heo /* stop DMA engine */ 2461fd7a697STejun Heo writeb(pp->dfl_prdctl, port_base + PORT_PRD_CTL); 2471fd7a697STejun Heo } 2481fd7a697STejun Heo 2491fd7a697STejun Heo static u8 inic_bmdma_status(struct ata_port *ap) 2501fd7a697STejun Heo { 2511fd7a697STejun Heo /* event is already verified by the interrupt handler */ 2521fd7a697STejun Heo return ATA_DMA_INTR; 2531fd7a697STejun Heo } 2541fd7a697STejun Heo 2551fd7a697STejun Heo static void inic_host_intr(struct ata_port *ap) 2561fd7a697STejun Heo { 2571fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 2589af5c9c9STejun Heo struct ata_eh_info *ehi = &ap->link.eh_info; 2591fd7a697STejun Heo u8 irq_stat; 2601fd7a697STejun Heo 2611fd7a697STejun Heo /* fetch and clear irq */ 2621fd7a697STejun Heo irq_stat = readb(port_base + PORT_IRQ_STAT); 2631fd7a697STejun Heo writeb(irq_stat, port_base + PORT_IRQ_STAT); 2641fd7a697STejun Heo 2651fd7a697STejun Heo if (likely(!(irq_stat & PIRQ_ERR))) { 2669af5c9c9STejun Heo struct ata_queued_cmd *qc = 2679af5c9c9STejun Heo ata_qc_from_tag(ap, ap->link.active_tag); 2681fd7a697STejun Heo 2691fd7a697STejun Heo if (unlikely(!qc || (qc->tf.flags & ATA_TFLAG_POLLING))) { 2706fd36390STejun Heo ap->ops->check_status(ap); /* clear ATA interrupt */ 2711fd7a697STejun Heo return; 2721fd7a697STejun Heo } 2731fd7a697STejun Heo 274*9363c382STejun Heo if (likely(ata_sff_host_intr(ap, qc))) 2751fd7a697STejun Heo return; 2761fd7a697STejun Heo 2776fd36390STejun Heo ap->ops->check_status(ap); /* clear ATA interrupt */ 2781fd7a697STejun Heo ata_port_printk(ap, KERN_WARNING, "unhandled " 2791fd7a697STejun Heo "interrupt, irq_stat=%x\n", irq_stat); 2801fd7a697STejun Heo return; 2811fd7a697STejun Heo } 2821fd7a697STejun Heo 2831fd7a697STejun Heo /* error */ 2841fd7a697STejun Heo ata_ehi_push_desc(ehi, "irq_stat=0x%x", irq_stat); 2851fd7a697STejun Heo 2861fd7a697STejun Heo if (irq_stat & (PIRQ_OFFLINE | PIRQ_ONLINE)) { 2871fd7a697STejun Heo ata_ehi_hotplugged(ehi); 2881fd7a697STejun Heo ata_port_freeze(ap); 2891fd7a697STejun Heo } else 2901fd7a697STejun Heo ata_port_abort(ap); 2911fd7a697STejun Heo } 2921fd7a697STejun Heo 2931fd7a697STejun Heo static irqreturn_t inic_interrupt(int irq, void *dev_instance) 2941fd7a697STejun Heo { 2951fd7a697STejun Heo struct ata_host *host = dev_instance; 2960d5ff566STejun Heo void __iomem *mmio_base = host->iomap[MMIO_BAR]; 2971fd7a697STejun Heo u16 host_irq_stat; 2981fd7a697STejun Heo int i, handled = 0;; 2991fd7a697STejun Heo 3001fd7a697STejun Heo host_irq_stat = readw(mmio_base + HOST_IRQ_STAT); 3011fd7a697STejun Heo 3021fd7a697STejun Heo if (unlikely(!(host_irq_stat & HIRQ_GLOBAL))) 3031fd7a697STejun Heo goto out; 3041fd7a697STejun Heo 3051fd7a697STejun Heo spin_lock(&host->lock); 3061fd7a697STejun Heo 3071fd7a697STejun Heo for (i = 0; i < NR_PORTS; i++) { 3081fd7a697STejun Heo struct ata_port *ap = host->ports[i]; 3091fd7a697STejun Heo 3101fd7a697STejun Heo if (!(host_irq_stat & (HIRQ_PORT0 << i))) 3111fd7a697STejun Heo continue; 3121fd7a697STejun Heo 3131fd7a697STejun Heo if (likely(ap && !(ap->flags & ATA_FLAG_DISABLED))) { 3141fd7a697STejun Heo inic_host_intr(ap); 3151fd7a697STejun Heo handled++; 3161fd7a697STejun Heo } else { 3171fd7a697STejun Heo if (ata_ratelimit()) 3181fd7a697STejun Heo dev_printk(KERN_ERR, host->dev, "interrupt " 3191fd7a697STejun Heo "from disabled port %d (0x%x)\n", 3201fd7a697STejun Heo i, host_irq_stat); 3211fd7a697STejun Heo } 3221fd7a697STejun Heo } 3231fd7a697STejun Heo 3241fd7a697STejun Heo spin_unlock(&host->lock); 3251fd7a697STejun Heo 3261fd7a697STejun Heo out: 3271fd7a697STejun Heo return IRQ_RETVAL(handled); 3281fd7a697STejun Heo } 3291fd7a697STejun Heo 3301fd7a697STejun Heo static unsigned int inic_qc_issue(struct ata_queued_cmd *qc) 3311fd7a697STejun Heo { 3321fd7a697STejun Heo struct ata_port *ap = qc->ap; 3331fd7a697STejun Heo 3341fd7a697STejun Heo /* ATA IRQ doesn't wait for DMA transfer completion and vice 3351fd7a697STejun Heo * versa. Mask IRQ selectively to detect command completion. 3361fd7a697STejun Heo * Without it, ATA DMA read command can cause data corruption. 3371fd7a697STejun Heo * 3381fd7a697STejun Heo * Something similar might be needed for ATAPI writes. I 3391fd7a697STejun Heo * tried a lot of combinations but couldn't find the solution. 3401fd7a697STejun Heo */ 3411fd7a697STejun Heo if (qc->tf.protocol == ATA_PROT_DMA && 3421fd7a697STejun Heo !(qc->tf.flags & ATA_TFLAG_WRITE)) 3431fd7a697STejun Heo inic_set_pirq_mask(ap, PIRQ_MASK_DMA_READ); 3441fd7a697STejun Heo else 3451fd7a697STejun Heo inic_set_pirq_mask(ap, PIRQ_MASK_OTHER); 3461fd7a697STejun Heo 3471fd7a697STejun Heo /* Issuing a command to yet uninitialized port locks up the 3481fd7a697STejun Heo * controller. Most of the time, this happens for the first 3491fd7a697STejun Heo * command after reset which are ATA and ATAPI IDENTIFYs. 3501fd7a697STejun Heo * Fast fail if stat is 0x7f or 0xff for those commands. 3511fd7a697STejun Heo */ 3521fd7a697STejun Heo if (unlikely(qc->tf.command == ATA_CMD_ID_ATA || 3531fd7a697STejun Heo qc->tf.command == ATA_CMD_ID_ATAPI)) { 3546fd36390STejun Heo u8 stat = ap->ops->check_status(ap); 3551fd7a697STejun Heo if (stat == 0x7f || stat == 0xff) 3561fd7a697STejun Heo return AC_ERR_HSM; 3571fd7a697STejun Heo } 3581fd7a697STejun Heo 359*9363c382STejun Heo return ata_sff_qc_issue(qc); 3601fd7a697STejun Heo } 3611fd7a697STejun Heo 3621fd7a697STejun Heo static void inic_freeze(struct ata_port *ap) 3631fd7a697STejun Heo { 3641fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 3651fd7a697STejun Heo 3661fd7a697STejun Heo __inic_set_pirq_mask(ap, PIRQ_MASK_FREEZE); 3671fd7a697STejun Heo 3686fd36390STejun Heo ap->ops->check_status(ap); 3691fd7a697STejun Heo writeb(0xff, port_base + PORT_IRQ_STAT); 3701fd7a697STejun Heo 3711fd7a697STejun Heo readb(port_base + PORT_IRQ_STAT); /* flush */ 3721fd7a697STejun Heo } 3731fd7a697STejun Heo 3741fd7a697STejun Heo static void inic_thaw(struct ata_port *ap) 3751fd7a697STejun Heo { 3761fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 3771fd7a697STejun Heo 3786fd36390STejun Heo ap->ops->check_status(ap); 3791fd7a697STejun Heo writeb(0xff, port_base + PORT_IRQ_STAT); 3801fd7a697STejun Heo 3811fd7a697STejun Heo __inic_set_pirq_mask(ap, PIRQ_MASK_OTHER); 3821fd7a697STejun Heo 3831fd7a697STejun Heo readb(port_base + PORT_IRQ_STAT); /* flush */ 3841fd7a697STejun Heo } 3851fd7a697STejun Heo 3861fd7a697STejun Heo /* 3871fd7a697STejun Heo * SRST and SControl hardreset don't give valid signature on this 3881fd7a697STejun Heo * controller. Only controller specific hardreset mechanism works. 3891fd7a697STejun Heo */ 390cc0680a5STejun Heo static int inic_hardreset(struct ata_link *link, unsigned int *class, 391d4b2bab4STejun Heo unsigned long deadline) 3921fd7a697STejun Heo { 393cc0680a5STejun Heo struct ata_port *ap = link->ap; 3941fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 3951fd7a697STejun Heo void __iomem *idma_ctl = port_base + PORT_IDMA_CTL; 396cc0680a5STejun Heo const unsigned long *timing = sata_ehc_deb_timing(&link->eh_context); 3971fd7a697STejun Heo u16 val; 3981fd7a697STejun Heo int rc; 3991fd7a697STejun Heo 4001fd7a697STejun Heo /* hammer it into sane state */ 4011fd7a697STejun Heo inic_reset_port(port_base); 4021fd7a697STejun Heo 4031fd7a697STejun Heo val = readw(idma_ctl); 4041fd7a697STejun Heo writew(val | IDMA_CTL_RST_ATA, idma_ctl); 4051fd7a697STejun Heo readw(idma_ctl); /* flush */ 4061fd7a697STejun Heo msleep(1); 4071fd7a697STejun Heo writew(val & ~IDMA_CTL_RST_ATA, idma_ctl); 4081fd7a697STejun Heo 409cc0680a5STejun Heo rc = sata_link_resume(link, timing, deadline); 4101fd7a697STejun Heo if (rc) { 411cc0680a5STejun Heo ata_link_printk(link, KERN_WARNING, "failed to resume " 412fe334602STejun Heo "link after reset (errno=%d)\n", rc); 4131fd7a697STejun Heo return rc; 4141fd7a697STejun Heo } 4151fd7a697STejun Heo 4161fd7a697STejun Heo *class = ATA_DEV_NONE; 417cc0680a5STejun Heo if (ata_link_online(link)) { 4181fd7a697STejun Heo struct ata_taskfile tf; 4191fd7a697STejun Heo 420fe334602STejun Heo /* wait a while before checking status */ 421*9363c382STejun Heo ata_sff_wait_after_reset(ap, deadline); 422fe334602STejun Heo 423*9363c382STejun Heo rc = ata_sff_wait_ready(ap, deadline); 4249b89391cSTejun Heo /* link occupied, -ENODEV too is an error */ 4259b89391cSTejun Heo if (rc) { 426cc0680a5STejun Heo ata_link_printk(link, KERN_WARNING, "device not ready " 427d4b2bab4STejun Heo "after hardreset (errno=%d)\n", rc); 428d4b2bab4STejun Heo return rc; 4291fd7a697STejun Heo } 4301fd7a697STejun Heo 431*9363c382STejun Heo ata_sff_tf_read(ap, &tf); 4321fd7a697STejun Heo *class = ata_dev_classify(&tf); 4331fd7a697STejun Heo if (*class == ATA_DEV_UNKNOWN) 4341fd7a697STejun Heo *class = ATA_DEV_NONE; 4351fd7a697STejun Heo } 4361fd7a697STejun Heo 4371fd7a697STejun Heo return 0; 4381fd7a697STejun Heo } 4391fd7a697STejun Heo 4401fd7a697STejun Heo static void inic_error_handler(struct ata_port *ap) 4411fd7a697STejun Heo { 4421fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 4431fd7a697STejun Heo struct inic_port_priv *pp = ap->private_data; 4441fd7a697STejun Heo unsigned long flags; 4451fd7a697STejun Heo 4461fd7a697STejun Heo /* reset PIO HSM and stop DMA engine */ 4471fd7a697STejun Heo inic_reset_port(port_base); 4481fd7a697STejun Heo 4491fd7a697STejun Heo spin_lock_irqsave(ap->lock, flags); 4501fd7a697STejun Heo ap->hsm_task_state = HSM_ST_IDLE; 4511fd7a697STejun Heo writeb(pp->dfl_prdctl, port_base + PORT_PRD_CTL); 4521fd7a697STejun Heo spin_unlock_irqrestore(ap->lock, flags); 4531fd7a697STejun Heo 4541fd7a697STejun Heo /* PIO and DMA engines have been stopped, perform recovery */ 455a1efdabaSTejun Heo ata_std_error_handler(ap); 4561fd7a697STejun Heo } 4571fd7a697STejun Heo 4581fd7a697STejun Heo static void inic_post_internal_cmd(struct ata_queued_cmd *qc) 4591fd7a697STejun Heo { 4601fd7a697STejun Heo /* make DMA engine forget about the failed command */ 461a51d644aSTejun Heo if (qc->flags & ATA_QCFLAG_FAILED) 4621fd7a697STejun Heo inic_reset_port(inic_port_base(qc->ap)); 4631fd7a697STejun Heo } 4641fd7a697STejun Heo 465cd0d3bbcSAlan static void inic_dev_config(struct ata_device *dev) 4661fd7a697STejun Heo { 4671fd7a697STejun Heo /* inic can only handle upto LBA28 max sectors */ 4681fd7a697STejun Heo if (dev->max_sectors > ATA_MAX_SECTORS) 4691fd7a697STejun Heo dev->max_sectors = ATA_MAX_SECTORS; 47090c93785STejun Heo 47190c93785STejun Heo if (dev->n_sectors >= 1 << 28) { 47290c93785STejun Heo ata_dev_printk(dev, KERN_ERR, 47390c93785STejun Heo "ERROR: This driver doesn't support LBA48 yet and may cause\n" 47490c93785STejun Heo " data corruption on such devices. Disabling.\n"); 47590c93785STejun Heo ata_dev_disable(dev); 47690c93785STejun Heo } 4771fd7a697STejun Heo } 4781fd7a697STejun Heo 4791fd7a697STejun Heo static void init_port(struct ata_port *ap) 4801fd7a697STejun Heo { 4811fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 4821fd7a697STejun Heo 4831fd7a697STejun Heo /* Setup PRD address */ 4841fd7a697STejun Heo writel(ap->prd_dma, port_base + PORT_PRD_ADDR); 4851fd7a697STejun Heo } 4861fd7a697STejun Heo 4871fd7a697STejun Heo static int inic_port_resume(struct ata_port *ap) 4881fd7a697STejun Heo { 4891fd7a697STejun Heo init_port(ap); 4901fd7a697STejun Heo return 0; 4911fd7a697STejun Heo } 4921fd7a697STejun Heo 4931fd7a697STejun Heo static int inic_port_start(struct ata_port *ap) 4941fd7a697STejun Heo { 4951fd7a697STejun Heo void __iomem *port_base = inic_port_base(ap); 4961fd7a697STejun Heo struct inic_port_priv *pp; 4971fd7a697STejun Heo u8 tmp; 4981fd7a697STejun Heo int rc; 4991fd7a697STejun Heo 5001fd7a697STejun Heo /* alloc and initialize private data */ 50124dc5f33STejun Heo pp = devm_kzalloc(ap->host->dev, sizeof(*pp), GFP_KERNEL); 5021fd7a697STejun Heo if (!pp) 5031fd7a697STejun Heo return -ENOMEM; 5041fd7a697STejun Heo ap->private_data = pp; 5051fd7a697STejun Heo 5061fd7a697STejun Heo /* default PRD_CTL value, DMAEN, WR and START off */ 5071fd7a697STejun Heo tmp = readb(port_base + PORT_PRD_CTL); 5081fd7a697STejun Heo tmp &= ~(PRD_CTL_DMAEN | PRD_CTL_WR | PRD_CTL_START); 5091fd7a697STejun Heo pp->dfl_prdctl = tmp; 5101fd7a697STejun Heo 5111fd7a697STejun Heo /* Alloc resources */ 5121fd7a697STejun Heo rc = ata_port_start(ap); 5131fd7a697STejun Heo if (rc) { 5141fd7a697STejun Heo kfree(pp); 5151fd7a697STejun Heo return rc; 5161fd7a697STejun Heo } 5171fd7a697STejun Heo 5181fd7a697STejun Heo init_port(ap); 5191fd7a697STejun Heo 5201fd7a697STejun Heo return 0; 5211fd7a697STejun Heo } 5221fd7a697STejun Heo 5231fd7a697STejun Heo static struct ata_port_operations inic_port_ops = { 524029cfd6bSTejun Heo .inherits = &ata_sff_port_ops, 5251fd7a697STejun Heo 5261fd7a697STejun Heo .bmdma_setup = inic_bmdma_setup, 5271fd7a697STejun Heo .bmdma_start = inic_bmdma_start, 5281fd7a697STejun Heo .bmdma_stop = inic_bmdma_stop, 5291fd7a697STejun Heo .bmdma_status = inic_bmdma_status, 5301fd7a697STejun Heo .qc_issue = inic_qc_issue, 5311fd7a697STejun Heo 5321fd7a697STejun Heo .freeze = inic_freeze, 5331fd7a697STejun Heo .thaw = inic_thaw, 534a1efdabaSTejun Heo .softreset = ATA_OP_NULL, /* softreset is broken */ 535a1efdabaSTejun Heo .hardreset = inic_hardreset, 5361fd7a697STejun Heo .error_handler = inic_error_handler, 5371fd7a697STejun Heo .post_internal_cmd = inic_post_internal_cmd, 5381fd7a697STejun Heo .dev_config = inic_dev_config, 5391fd7a697STejun Heo 540029cfd6bSTejun Heo .scr_read = inic_scr_read, 541029cfd6bSTejun Heo .scr_write = inic_scr_write, 5421fd7a697STejun Heo 543029cfd6bSTejun Heo .port_resume = inic_port_resume, 5441fd7a697STejun Heo .port_start = inic_port_start, 5451fd7a697STejun Heo }; 5461fd7a697STejun Heo 5471fd7a697STejun Heo static struct ata_port_info inic_port_info = { 5480dc36888STejun Heo /* For some reason, ATAPI_PROT_PIO is broken on this 5491fd7a697STejun Heo * controller, and no, PIO_POLLING does't fix it. It somehow 5501fd7a697STejun Heo * manages to report the wrong ireason and ignoring ireason 5511fd7a697STejun Heo * results in machine lock up. Tell libata to always prefer 5521fd7a697STejun Heo * DMA. 5531fd7a697STejun Heo */ 5541fd7a697STejun Heo .flags = ATA_FLAG_SATA | ATA_FLAG_PIO_DMA, 5551fd7a697STejun Heo .pio_mask = 0x1f, /* pio0-4 */ 5561fd7a697STejun Heo .mwdma_mask = 0x07, /* mwdma0-2 */ 557bf6263a8SJeff Garzik .udma_mask = ATA_UDMA6, 5581fd7a697STejun Heo .port_ops = &inic_port_ops 5591fd7a697STejun Heo }; 5601fd7a697STejun Heo 5611fd7a697STejun Heo static int init_controller(void __iomem *mmio_base, u16 hctl) 5621fd7a697STejun Heo { 5631fd7a697STejun Heo int i; 5641fd7a697STejun Heo u16 val; 5651fd7a697STejun Heo 5661fd7a697STejun Heo hctl &= ~HCTL_KNOWN_BITS; 5671fd7a697STejun Heo 5681fd7a697STejun Heo /* Soft reset whole controller. Spec says reset duration is 3 5691fd7a697STejun Heo * PCI clocks, be generous and give it 10ms. 5701fd7a697STejun Heo */ 5711fd7a697STejun Heo writew(hctl | HCTL_SOFTRST, mmio_base + HOST_CTL); 5721fd7a697STejun Heo readw(mmio_base + HOST_CTL); /* flush */ 5731fd7a697STejun Heo 5741fd7a697STejun Heo for (i = 0; i < 10; i++) { 5751fd7a697STejun Heo msleep(1); 5761fd7a697STejun Heo val = readw(mmio_base + HOST_CTL); 5771fd7a697STejun Heo if (!(val & HCTL_SOFTRST)) 5781fd7a697STejun Heo break; 5791fd7a697STejun Heo } 5801fd7a697STejun Heo 5811fd7a697STejun Heo if (val & HCTL_SOFTRST) 5821fd7a697STejun Heo return -EIO; 5831fd7a697STejun Heo 5841fd7a697STejun Heo /* mask all interrupts and reset ports */ 5851fd7a697STejun Heo for (i = 0; i < NR_PORTS; i++) { 5861fd7a697STejun Heo void __iomem *port_base = mmio_base + i * PORT_SIZE; 5871fd7a697STejun Heo 5881fd7a697STejun Heo writeb(0xff, port_base + PORT_IRQ_MASK); 5891fd7a697STejun Heo inic_reset_port(port_base); 5901fd7a697STejun Heo } 5911fd7a697STejun Heo 5921fd7a697STejun Heo /* port IRQ is masked now, unmask global IRQ */ 5931fd7a697STejun Heo writew(hctl & ~HCTL_IRQOFF, mmio_base + HOST_CTL); 5941fd7a697STejun Heo val = readw(mmio_base + HOST_IRQ_MASK); 5951fd7a697STejun Heo val &= ~(HIRQ_PORT0 | HIRQ_PORT1); 5961fd7a697STejun Heo writew(val, mmio_base + HOST_IRQ_MASK); 5971fd7a697STejun Heo 5981fd7a697STejun Heo return 0; 5991fd7a697STejun Heo } 6001fd7a697STejun Heo 601438ac6d5STejun Heo #ifdef CONFIG_PM 6021fd7a697STejun Heo static int inic_pci_device_resume(struct pci_dev *pdev) 6031fd7a697STejun Heo { 6041fd7a697STejun Heo struct ata_host *host = dev_get_drvdata(&pdev->dev); 6051fd7a697STejun Heo struct inic_host_priv *hpriv = host->private_data; 6060d5ff566STejun Heo void __iomem *mmio_base = host->iomap[MMIO_BAR]; 6071fd7a697STejun Heo int rc; 6081fd7a697STejun Heo 6095aea408dSDmitriy Monakhov rc = ata_pci_device_do_resume(pdev); 6105aea408dSDmitriy Monakhov if (rc) 6115aea408dSDmitriy Monakhov return rc; 6121fd7a697STejun Heo 6131fd7a697STejun Heo if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) { 6141fd7a697STejun Heo rc = init_controller(mmio_base, hpriv->cached_hctl); 6151fd7a697STejun Heo if (rc) 6161fd7a697STejun Heo return rc; 6171fd7a697STejun Heo } 6181fd7a697STejun Heo 6191fd7a697STejun Heo ata_host_resume(host); 6201fd7a697STejun Heo 6211fd7a697STejun Heo return 0; 6221fd7a697STejun Heo } 623438ac6d5STejun Heo #endif 6241fd7a697STejun Heo 6251fd7a697STejun Heo static int inic_init_one(struct pci_dev *pdev, const struct pci_device_id *ent) 6261fd7a697STejun Heo { 6271fd7a697STejun Heo static int printed_version; 6284447d351STejun Heo const struct ata_port_info *ppi[] = { &inic_port_info, NULL }; 6294447d351STejun Heo struct ata_host *host; 6301fd7a697STejun Heo struct inic_host_priv *hpriv; 6310d5ff566STejun Heo void __iomem * const *iomap; 6321fd7a697STejun Heo int i, rc; 6331fd7a697STejun Heo 6341fd7a697STejun Heo if (!printed_version++) 6351fd7a697STejun Heo dev_printk(KERN_DEBUG, &pdev->dev, "version " DRV_VERSION "\n"); 6361fd7a697STejun Heo 6374447d351STejun Heo /* alloc host */ 6384447d351STejun Heo host = ata_host_alloc_pinfo(&pdev->dev, ppi, NR_PORTS); 6394447d351STejun Heo hpriv = devm_kzalloc(&pdev->dev, sizeof(*hpriv), GFP_KERNEL); 6404447d351STejun Heo if (!host || !hpriv) 6414447d351STejun Heo return -ENOMEM; 6424447d351STejun Heo 6434447d351STejun Heo host->private_data = hpriv; 6444447d351STejun Heo 6454447d351STejun Heo /* acquire resources and fill host */ 64624dc5f33STejun Heo rc = pcim_enable_device(pdev); 6471fd7a697STejun Heo if (rc) 6481fd7a697STejun Heo return rc; 6491fd7a697STejun Heo 6500d5ff566STejun Heo rc = pcim_iomap_regions(pdev, 0x3f, DRV_NAME); 6510d5ff566STejun Heo if (rc) 6520d5ff566STejun Heo return rc; 6534447d351STejun Heo host->iomap = iomap = pcim_iomap_table(pdev); 6544447d351STejun Heo 6554447d351STejun Heo for (i = 0; i < NR_PORTS; i++) { 656cbcdd875STejun Heo struct ata_port *ap = host->ports[i]; 657cbcdd875STejun Heo struct ata_ioports *port = &ap->ioaddr; 658cbcdd875STejun Heo unsigned int offset = i * PORT_SIZE; 6594447d351STejun Heo 6604447d351STejun Heo port->cmd_addr = iomap[2 * i]; 6614447d351STejun Heo port->altstatus_addr = 6624447d351STejun Heo port->ctl_addr = (void __iomem *) 6634447d351STejun Heo ((unsigned long)iomap[2 * i + 1] | ATA_PCI_CTL_OFS); 664cbcdd875STejun Heo port->scr_addr = iomap[MMIO_BAR] + offset + PORT_SCR; 6654447d351STejun Heo 666*9363c382STejun Heo ata_sff_std_ports(port); 667cbcdd875STejun Heo 668cbcdd875STejun Heo ata_port_pbar_desc(ap, MMIO_BAR, -1, "mmio"); 669cbcdd875STejun Heo ata_port_pbar_desc(ap, MMIO_BAR, offset, "port"); 670cbcdd875STejun Heo ata_port_desc(ap, "cmd 0x%llx ctl 0x%llx", 671cbcdd875STejun Heo (unsigned long long)pci_resource_start(pdev, 2 * i), 672cbcdd875STejun Heo (unsigned long long)pci_resource_start(pdev, (2 * i + 1)) | 673cbcdd875STejun Heo ATA_PCI_CTL_OFS); 6744447d351STejun Heo } 6754447d351STejun Heo 6764447d351STejun Heo hpriv->cached_hctl = readw(iomap[MMIO_BAR] + HOST_CTL); 6771fd7a697STejun Heo 6781fd7a697STejun Heo /* Set dma_mask. This devices doesn't support 64bit addressing. */ 6791fd7a697STejun Heo rc = pci_set_dma_mask(pdev, DMA_32BIT_MASK); 6801fd7a697STejun Heo if (rc) { 6811fd7a697STejun Heo dev_printk(KERN_ERR, &pdev->dev, 6821fd7a697STejun Heo "32-bit DMA enable failed\n"); 68324dc5f33STejun Heo return rc; 6841fd7a697STejun Heo } 6851fd7a697STejun Heo 6861fd7a697STejun Heo rc = pci_set_consistent_dma_mask(pdev, DMA_32BIT_MASK); 6871fd7a697STejun Heo if (rc) { 6881fd7a697STejun Heo dev_printk(KERN_ERR, &pdev->dev, 6891fd7a697STejun Heo "32-bit consistent DMA enable failed\n"); 69024dc5f33STejun Heo return rc; 6911fd7a697STejun Heo } 6921fd7a697STejun Heo 693b7d8629fSFUJITA Tomonori /* 694b7d8629fSFUJITA Tomonori * This controller is braindamaged. dma_boundary is 0xffff 695b7d8629fSFUJITA Tomonori * like others but it will lock up the whole machine HARD if 696b7d8629fSFUJITA Tomonori * 65536 byte PRD entry is fed. Reduce maximum segment size. 697b7d8629fSFUJITA Tomonori */ 698b7d8629fSFUJITA Tomonori rc = pci_set_dma_max_seg_size(pdev, 65536 - 512); 699b7d8629fSFUJITA Tomonori if (rc) { 700b7d8629fSFUJITA Tomonori dev_printk(KERN_ERR, &pdev->dev, 701b7d8629fSFUJITA Tomonori "failed to set the maximum segment size.\n"); 702b7d8629fSFUJITA Tomonori return rc; 703b7d8629fSFUJITA Tomonori } 704b7d8629fSFUJITA Tomonori 7050d5ff566STejun Heo rc = init_controller(iomap[MMIO_BAR], hpriv->cached_hctl); 7061fd7a697STejun Heo if (rc) { 7071fd7a697STejun Heo dev_printk(KERN_ERR, &pdev->dev, 7081fd7a697STejun Heo "failed to initialize controller\n"); 70924dc5f33STejun Heo return rc; 7101fd7a697STejun Heo } 7111fd7a697STejun Heo 7121fd7a697STejun Heo pci_set_master(pdev); 7134447d351STejun Heo return ata_host_activate(host, pdev->irq, inic_interrupt, IRQF_SHARED, 7144447d351STejun Heo &inic_sht); 7151fd7a697STejun Heo } 7161fd7a697STejun Heo 7171fd7a697STejun Heo static const struct pci_device_id inic_pci_tbl[] = { 7181fd7a697STejun Heo { PCI_VDEVICE(INIT, 0x1622), }, 7191fd7a697STejun Heo { }, 7201fd7a697STejun Heo }; 7211fd7a697STejun Heo 7221fd7a697STejun Heo static struct pci_driver inic_pci_driver = { 7231fd7a697STejun Heo .name = DRV_NAME, 7241fd7a697STejun Heo .id_table = inic_pci_tbl, 725438ac6d5STejun Heo #ifdef CONFIG_PM 7261fd7a697STejun Heo .suspend = ata_pci_device_suspend, 7271fd7a697STejun Heo .resume = inic_pci_device_resume, 728438ac6d5STejun Heo #endif 7291fd7a697STejun Heo .probe = inic_init_one, 7301fd7a697STejun Heo .remove = ata_pci_remove_one, 7311fd7a697STejun Heo }; 7321fd7a697STejun Heo 7331fd7a697STejun Heo static int __init inic_init(void) 7341fd7a697STejun Heo { 7351fd7a697STejun Heo return pci_register_driver(&inic_pci_driver); 7361fd7a697STejun Heo } 7371fd7a697STejun Heo 7381fd7a697STejun Heo static void __exit inic_exit(void) 7391fd7a697STejun Heo { 7401fd7a697STejun Heo pci_unregister_driver(&inic_pci_driver); 7411fd7a697STejun Heo } 7421fd7a697STejun Heo 7431fd7a697STejun Heo MODULE_AUTHOR("Tejun Heo"); 7441fd7a697STejun Heo MODULE_DESCRIPTION("low-level driver for Initio 162x SATA"); 7451fd7a697STejun Heo MODULE_LICENSE("GPL v2"); 7461fd7a697STejun Heo MODULE_DEVICE_TABLE(pci, inic_pci_tbl); 7471fd7a697STejun Heo MODULE_VERSION(DRV_VERSION); 7481fd7a697STejun Heo 7491fd7a697STejun Heo module_init(inic_init); 7501fd7a697STejun Heo module_exit(inic_exit); 751