1c6fd2807SJeff Garzik /* 2c6fd2807SJeff Garzik * libata-core.c - helper library for ATA 3c6fd2807SJeff Garzik * 4c6fd2807SJeff Garzik * Maintained by: Jeff Garzik <jgarzik@pobox.com> 5c6fd2807SJeff Garzik * Please ALWAYS copy linux-ide@vger.kernel.org 6c6fd2807SJeff Garzik * on emails. 7c6fd2807SJeff Garzik * 8c6fd2807SJeff Garzik * Copyright 2003-2004 Red Hat, Inc. All rights reserved. 9c6fd2807SJeff Garzik * Copyright 2003-2004 Jeff Garzik 10c6fd2807SJeff Garzik * 11c6fd2807SJeff Garzik * 12c6fd2807SJeff Garzik * This program is free software; you can redistribute it and/or modify 13c6fd2807SJeff Garzik * it under the terms of the GNU General Public License as published by 14c6fd2807SJeff Garzik * the Free Software Foundation; either version 2, or (at your option) 15c6fd2807SJeff Garzik * any later version. 16c6fd2807SJeff Garzik * 17c6fd2807SJeff Garzik * This program is distributed in the hope that it will be useful, 18c6fd2807SJeff Garzik * but WITHOUT ANY WARRANTY; without even the implied warranty of 19c6fd2807SJeff Garzik * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the 20c6fd2807SJeff Garzik * GNU General Public License for more details. 21c6fd2807SJeff Garzik * 22c6fd2807SJeff Garzik * You should have received a copy of the GNU General Public License 23c6fd2807SJeff Garzik * along with this program; see the file COPYING. If not, write to 24c6fd2807SJeff Garzik * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. 25c6fd2807SJeff Garzik * 26c6fd2807SJeff Garzik * 27c6fd2807SJeff Garzik * libata documentation is available via 'make {ps|pdf}docs', 28c6fd2807SJeff Garzik * as Documentation/DocBook/libata.* 29c6fd2807SJeff Garzik * 30c6fd2807SJeff Garzik * Hardware documentation available from http://www.t13.org/ and 31c6fd2807SJeff Garzik * http://www.sata-io.org/ 32c6fd2807SJeff Garzik * 3392c52c52SAlan Cox * Standards documents from: 3492c52c52SAlan Cox * http://www.t13.org (ATA standards, PCI DMA IDE spec) 3592c52c52SAlan Cox * http://www.t10.org (SCSI MMC - for ATAPI MMC) 3692c52c52SAlan Cox * http://www.sata-io.org (SATA) 3792c52c52SAlan Cox * http://www.compactflash.org (CF) 3892c52c52SAlan Cox * http://www.qic.org (QIC157 - Tape and DSC) 3992c52c52SAlan Cox * http://www.ce-ata.org (CE-ATA: not supported) 4092c52c52SAlan Cox * 41c6fd2807SJeff Garzik */ 42c6fd2807SJeff Garzik 43c6fd2807SJeff Garzik #include <linux/kernel.h> 44c6fd2807SJeff Garzik #include <linux/module.h> 45c6fd2807SJeff Garzik #include <linux/pci.h> 46c6fd2807SJeff Garzik #include <linux/init.h> 47c6fd2807SJeff Garzik #include <linux/list.h> 48c6fd2807SJeff Garzik #include <linux/mm.h> 49c6fd2807SJeff Garzik #include <linux/highmem.h> 50c6fd2807SJeff Garzik #include <linux/spinlock.h> 51c6fd2807SJeff Garzik #include <linux/blkdev.h> 52c6fd2807SJeff Garzik #include <linux/delay.h> 53c6fd2807SJeff Garzik #include <linux/timer.h> 54c6fd2807SJeff Garzik #include <linux/interrupt.h> 55c6fd2807SJeff Garzik #include <linux/completion.h> 56c6fd2807SJeff Garzik #include <linux/suspend.h> 57c6fd2807SJeff Garzik #include <linux/workqueue.h> 58c6fd2807SJeff Garzik #include <linux/jiffies.h> 59c6fd2807SJeff Garzik #include <linux/scatterlist.h> 602dcb407eSJeff Garzik #include <linux/io.h> 61c6fd2807SJeff Garzik #include <scsi/scsi.h> 62c6fd2807SJeff Garzik #include <scsi/scsi_cmnd.h> 63c6fd2807SJeff Garzik #include <scsi/scsi_host.h> 64c6fd2807SJeff Garzik #include <linux/libata.h> 65c6fd2807SJeff Garzik #include <asm/semaphore.h> 66c6fd2807SJeff Garzik #include <asm/byteorder.h> 67140b5e59STejun Heo #include <linux/cdrom.h> 68c6fd2807SJeff Garzik 69c6fd2807SJeff Garzik #include "libata.h" 70c6fd2807SJeff Garzik 71fda0efc5SJeff Garzik 72c6fd2807SJeff Garzik /* debounce timing parameters in msecs { interval, duration, timeout } */ 73c6fd2807SJeff Garzik const unsigned long sata_deb_timing_normal[] = { 5, 100, 2000 }; 74c6fd2807SJeff Garzik const unsigned long sata_deb_timing_hotplug[] = { 25, 500, 2000 }; 75c6fd2807SJeff Garzik const unsigned long sata_deb_timing_long[] = { 100, 2000, 5000 }; 76c6fd2807SJeff Garzik 77c6fd2807SJeff Garzik static unsigned int ata_dev_init_params(struct ata_device *dev, 78c6fd2807SJeff Garzik u16 heads, u16 sectors); 79c6fd2807SJeff Garzik static unsigned int ata_dev_set_xfermode(struct ata_device *dev); 80218f3d30SJeff Garzik static unsigned int ata_dev_set_feature(struct ata_device *dev, 81218f3d30SJeff Garzik u8 enable, u8 feature); 82c6fd2807SJeff Garzik static void ata_dev_xfermask(struct ata_device *dev); 8375683fe7STejun Heo static unsigned long ata_dev_blacklisted(const struct ata_device *dev); 84c6fd2807SJeff Garzik 85f3187195STejun Heo unsigned int ata_print_id = 1; 86c6fd2807SJeff Garzik static struct workqueue_struct *ata_wq; 87c6fd2807SJeff Garzik 88c6fd2807SJeff Garzik struct workqueue_struct *ata_aux_wq; 89c6fd2807SJeff Garzik 9033267325STejun Heo struct ata_force_param { 9133267325STejun Heo const char *name; 9233267325STejun Heo unsigned int cbl; 9333267325STejun Heo int spd_limit; 9433267325STejun Heo unsigned long xfer_mask; 9533267325STejun Heo unsigned int horkage_on; 9633267325STejun Heo unsigned int horkage_off; 9733267325STejun Heo }; 9833267325STejun Heo 9933267325STejun Heo struct ata_force_ent { 10033267325STejun Heo int port; 10133267325STejun Heo int device; 10233267325STejun Heo struct ata_force_param param; 10333267325STejun Heo }; 10433267325STejun Heo 10533267325STejun Heo static struct ata_force_ent *ata_force_tbl; 10633267325STejun Heo static int ata_force_tbl_size; 10733267325STejun Heo 10833267325STejun Heo static char ata_force_param_buf[PAGE_SIZE] __initdata; 1097afb4222STejun Heo /* param_buf is thrown away after initialization, disallow read */ 1107afb4222STejun Heo module_param_string(force, ata_force_param_buf, sizeof(ata_force_param_buf), 0); 11133267325STejun Heo MODULE_PARM_DESC(force, "Force ATA configurations including cable type, link speed and transfer mode (see Documentation/kernel-parameters.txt for details)"); 11233267325STejun Heo 113c6fd2807SJeff Garzik int atapi_enabled = 1; 114c6fd2807SJeff Garzik module_param(atapi_enabled, int, 0444); 115c6fd2807SJeff Garzik MODULE_PARM_DESC(atapi_enabled, "Enable discovery of ATAPI devices (0=off, 1=on)"); 116c6fd2807SJeff Garzik 117c5c61bdaSAdrian Bunk static int atapi_dmadir = 0; 118c6fd2807SJeff Garzik module_param(atapi_dmadir, int, 0444); 119c6fd2807SJeff Garzik MODULE_PARM_DESC(atapi_dmadir, "Enable ATAPI DMADIR bridge support (0=off, 1=on)"); 120c6fd2807SJeff Garzik 121baf4fdfaSMark Lord int atapi_passthru16 = 1; 122baf4fdfaSMark Lord module_param(atapi_passthru16, int, 0444); 123baf4fdfaSMark Lord MODULE_PARM_DESC(atapi_passthru16, "Enable ATA_16 passthru for ATAPI devices; on by default (0=off, 1=on)"); 124baf4fdfaSMark Lord 125c6fd2807SJeff Garzik int libata_fua = 0; 126c6fd2807SJeff Garzik module_param_named(fua, libata_fua, int, 0444); 127c6fd2807SJeff Garzik MODULE_PARM_DESC(fua, "FUA support (0=off, 1=on)"); 128c6fd2807SJeff Garzik 1292dcb407eSJeff Garzik static int ata_ignore_hpa; 1301e999736SAlan Cox module_param_named(ignore_hpa, ata_ignore_hpa, int, 0644); 1311e999736SAlan Cox MODULE_PARM_DESC(ignore_hpa, "Ignore HPA limit (0=keep BIOS limits, 1=ignore limits, using full disk)"); 1321e999736SAlan Cox 133b3a70601SAlan Cox static int libata_dma_mask = ATA_DMA_MASK_ATA|ATA_DMA_MASK_ATAPI|ATA_DMA_MASK_CFA; 134b3a70601SAlan Cox module_param_named(dma, libata_dma_mask, int, 0444); 135b3a70601SAlan Cox MODULE_PARM_DESC(dma, "DMA enable/disable (0x1==ATA, 0x2==ATAPI, 0x4==CF)"); 136b3a70601SAlan Cox 137c6fd2807SJeff Garzik static int ata_probe_timeout = ATA_TMOUT_INTERNAL / HZ; 138c6fd2807SJeff Garzik module_param(ata_probe_timeout, int, 0444); 139c6fd2807SJeff Garzik MODULE_PARM_DESC(ata_probe_timeout, "Set ATA probing timeout (seconds)"); 140c6fd2807SJeff Garzik 1416ebe9d86SJeff Garzik int libata_noacpi = 0; 142d7d0dad6SJeff Garzik module_param_named(noacpi, libata_noacpi, int, 0444); 1436ebe9d86SJeff Garzik MODULE_PARM_DESC(noacpi, "Disables the use of ACPI in probe/suspend/resume when set"); 14411ef697bSKristen Carlson Accardi 145ae8d4ee7SAlan Cox int libata_allow_tpm = 0; 146ae8d4ee7SAlan Cox module_param_named(allow_tpm, libata_allow_tpm, int, 0444); 147ae8d4ee7SAlan Cox MODULE_PARM_DESC(allow_tpm, "Permit the use of TPM commands"); 148ae8d4ee7SAlan Cox 149c6fd2807SJeff Garzik MODULE_AUTHOR("Jeff Garzik"); 150c6fd2807SJeff Garzik MODULE_DESCRIPTION("Library module for ATA devices"); 151c6fd2807SJeff Garzik MODULE_LICENSE("GPL"); 152c6fd2807SJeff Garzik MODULE_VERSION(DRV_VERSION); 153c6fd2807SJeff Garzik 154c6fd2807SJeff Garzik 155c6fd2807SJeff Garzik /** 15633267325STejun Heo * ata_force_cbl - force cable type according to libata.force 1574cdfa1b3SRandy Dunlap * @ap: ATA port of interest 15833267325STejun Heo * 15933267325STejun Heo * Force cable type according to libata.force and whine about it. 16033267325STejun Heo * The last entry which has matching port number is used, so it 16133267325STejun Heo * can be specified as part of device force parameters. For 16233267325STejun Heo * example, both "a:40c,1.00:udma4" and "1.00:40c,udma4" have the 16333267325STejun Heo * same effect. 16433267325STejun Heo * 16533267325STejun Heo * LOCKING: 16633267325STejun Heo * EH context. 16733267325STejun Heo */ 16833267325STejun Heo void ata_force_cbl(struct ata_port *ap) 16933267325STejun Heo { 17033267325STejun Heo int i; 17133267325STejun Heo 17233267325STejun Heo for (i = ata_force_tbl_size - 1; i >= 0; i--) { 17333267325STejun Heo const struct ata_force_ent *fe = &ata_force_tbl[i]; 17433267325STejun Heo 17533267325STejun Heo if (fe->port != -1 && fe->port != ap->print_id) 17633267325STejun Heo continue; 17733267325STejun Heo 17833267325STejun Heo if (fe->param.cbl == ATA_CBL_NONE) 17933267325STejun Heo continue; 18033267325STejun Heo 18133267325STejun Heo ap->cbl = fe->param.cbl; 18233267325STejun Heo ata_port_printk(ap, KERN_NOTICE, 18333267325STejun Heo "FORCE: cable set to %s\n", fe->param.name); 18433267325STejun Heo return; 18533267325STejun Heo } 18633267325STejun Heo } 18733267325STejun Heo 18833267325STejun Heo /** 18933267325STejun Heo * ata_force_spd_limit - force SATA spd limit according to libata.force 19033267325STejun Heo * @link: ATA link of interest 19133267325STejun Heo * 19233267325STejun Heo * Force SATA spd limit according to libata.force and whine about 19333267325STejun Heo * it. When only the port part is specified (e.g. 1:), the limit 19433267325STejun Heo * applies to all links connected to both the host link and all 19533267325STejun Heo * fan-out ports connected via PMP. If the device part is 19633267325STejun Heo * specified as 0 (e.g. 1.00:), it specifies the first fan-out 19733267325STejun Heo * link not the host link. Device number 15 always points to the 19833267325STejun Heo * host link whether PMP is attached or not. 19933267325STejun Heo * 20033267325STejun Heo * LOCKING: 20133267325STejun Heo * EH context. 20233267325STejun Heo */ 20333267325STejun Heo static void ata_force_spd_limit(struct ata_link *link) 20433267325STejun Heo { 20533267325STejun Heo int linkno, i; 20633267325STejun Heo 20733267325STejun Heo if (ata_is_host_link(link)) 20833267325STejun Heo linkno = 15; 20933267325STejun Heo else 21033267325STejun Heo linkno = link->pmp; 21133267325STejun Heo 21233267325STejun Heo for (i = ata_force_tbl_size - 1; i >= 0; i--) { 21333267325STejun Heo const struct ata_force_ent *fe = &ata_force_tbl[i]; 21433267325STejun Heo 21533267325STejun Heo if (fe->port != -1 && fe->port != link->ap->print_id) 21633267325STejun Heo continue; 21733267325STejun Heo 21833267325STejun Heo if (fe->device != -1 && fe->device != linkno) 21933267325STejun Heo continue; 22033267325STejun Heo 22133267325STejun Heo if (!fe->param.spd_limit) 22233267325STejun Heo continue; 22333267325STejun Heo 22433267325STejun Heo link->hw_sata_spd_limit = (1 << fe->param.spd_limit) - 1; 22533267325STejun Heo ata_link_printk(link, KERN_NOTICE, 22633267325STejun Heo "FORCE: PHY spd limit set to %s\n", fe->param.name); 22733267325STejun Heo return; 22833267325STejun Heo } 22933267325STejun Heo } 23033267325STejun Heo 23133267325STejun Heo /** 23233267325STejun Heo * ata_force_xfermask - force xfermask according to libata.force 23333267325STejun Heo * @dev: ATA device of interest 23433267325STejun Heo * 23533267325STejun Heo * Force xfer_mask according to libata.force and whine about it. 23633267325STejun Heo * For consistency with link selection, device number 15 selects 23733267325STejun Heo * the first device connected to the host link. 23833267325STejun Heo * 23933267325STejun Heo * LOCKING: 24033267325STejun Heo * EH context. 24133267325STejun Heo */ 24233267325STejun Heo static void ata_force_xfermask(struct ata_device *dev) 24333267325STejun Heo { 24433267325STejun Heo int devno = dev->link->pmp + dev->devno; 24533267325STejun Heo int alt_devno = devno; 24633267325STejun Heo int i; 24733267325STejun Heo 24833267325STejun Heo /* allow n.15 for the first device attached to host port */ 24933267325STejun Heo if (ata_is_host_link(dev->link) && devno == 0) 25033267325STejun Heo alt_devno = 15; 25133267325STejun Heo 25233267325STejun Heo for (i = ata_force_tbl_size - 1; i >= 0; i--) { 25333267325STejun Heo const struct ata_force_ent *fe = &ata_force_tbl[i]; 25433267325STejun Heo unsigned long pio_mask, mwdma_mask, udma_mask; 25533267325STejun Heo 25633267325STejun Heo if (fe->port != -1 && fe->port != dev->link->ap->print_id) 25733267325STejun Heo continue; 25833267325STejun Heo 25933267325STejun Heo if (fe->device != -1 && fe->device != devno && 26033267325STejun Heo fe->device != alt_devno) 26133267325STejun Heo continue; 26233267325STejun Heo 26333267325STejun Heo if (!fe->param.xfer_mask) 26433267325STejun Heo continue; 26533267325STejun Heo 26633267325STejun Heo ata_unpack_xfermask(fe->param.xfer_mask, 26733267325STejun Heo &pio_mask, &mwdma_mask, &udma_mask); 26833267325STejun Heo if (udma_mask) 26933267325STejun Heo dev->udma_mask = udma_mask; 27033267325STejun Heo else if (mwdma_mask) { 27133267325STejun Heo dev->udma_mask = 0; 27233267325STejun Heo dev->mwdma_mask = mwdma_mask; 27333267325STejun Heo } else { 27433267325STejun Heo dev->udma_mask = 0; 27533267325STejun Heo dev->mwdma_mask = 0; 27633267325STejun Heo dev->pio_mask = pio_mask; 27733267325STejun Heo } 27833267325STejun Heo 27933267325STejun Heo ata_dev_printk(dev, KERN_NOTICE, 28033267325STejun Heo "FORCE: xfer_mask set to %s\n", fe->param.name); 28133267325STejun Heo return; 28233267325STejun Heo } 28333267325STejun Heo } 28433267325STejun Heo 28533267325STejun Heo /** 28633267325STejun Heo * ata_force_horkage - force horkage according to libata.force 28733267325STejun Heo * @dev: ATA device of interest 28833267325STejun Heo * 28933267325STejun Heo * Force horkage according to libata.force and whine about it. 29033267325STejun Heo * For consistency with link selection, device number 15 selects 29133267325STejun Heo * the first device connected to the host link. 29233267325STejun Heo * 29333267325STejun Heo * LOCKING: 29433267325STejun Heo * EH context. 29533267325STejun Heo */ 29633267325STejun Heo static void ata_force_horkage(struct ata_device *dev) 29733267325STejun Heo { 29833267325STejun Heo int devno = dev->link->pmp + dev->devno; 29933267325STejun Heo int alt_devno = devno; 30033267325STejun Heo int i; 30133267325STejun Heo 30233267325STejun Heo /* allow n.15 for the first device attached to host port */ 30333267325STejun Heo if (ata_is_host_link(dev->link) && devno == 0) 30433267325STejun Heo alt_devno = 15; 30533267325STejun Heo 30633267325STejun Heo for (i = 0; i < ata_force_tbl_size; i++) { 30733267325STejun Heo const struct ata_force_ent *fe = &ata_force_tbl[i]; 30833267325STejun Heo 30933267325STejun Heo if (fe->port != -1 && fe->port != dev->link->ap->print_id) 31033267325STejun Heo continue; 31133267325STejun Heo 31233267325STejun Heo if (fe->device != -1 && fe->device != devno && 31333267325STejun Heo fe->device != alt_devno) 31433267325STejun Heo continue; 31533267325STejun Heo 31633267325STejun Heo if (!(~dev->horkage & fe->param.horkage_on) && 31733267325STejun Heo !(dev->horkage & fe->param.horkage_off)) 31833267325STejun Heo continue; 31933267325STejun Heo 32033267325STejun Heo dev->horkage |= fe->param.horkage_on; 32133267325STejun Heo dev->horkage &= ~fe->param.horkage_off; 32233267325STejun Heo 32333267325STejun Heo ata_dev_printk(dev, KERN_NOTICE, 32433267325STejun Heo "FORCE: horkage modified (%s)\n", fe->param.name); 32533267325STejun Heo } 32633267325STejun Heo } 32733267325STejun Heo 32833267325STejun Heo /** 329c6fd2807SJeff Garzik * ata_tf_to_fis - Convert ATA taskfile to SATA FIS structure 330c6fd2807SJeff Garzik * @tf: Taskfile to convert 331c6fd2807SJeff Garzik * @pmp: Port multiplier port 3329977126cSTejun Heo * @is_cmd: This FIS is for command 3339977126cSTejun Heo * @fis: Buffer into which data will output 334c6fd2807SJeff Garzik * 335c6fd2807SJeff Garzik * Converts a standard ATA taskfile to a Serial ATA 336c6fd2807SJeff Garzik * FIS structure (Register - Host to Device). 337c6fd2807SJeff Garzik * 338c6fd2807SJeff Garzik * LOCKING: 339c6fd2807SJeff Garzik * Inherited from caller. 340c6fd2807SJeff Garzik */ 3419977126cSTejun Heo void ata_tf_to_fis(const struct ata_taskfile *tf, u8 pmp, int is_cmd, u8 *fis) 342c6fd2807SJeff Garzik { 343c6fd2807SJeff Garzik fis[0] = 0x27; /* Register - Host to Device FIS */ 3449977126cSTejun Heo fis[1] = pmp & 0xf; /* Port multiplier number*/ 3459977126cSTejun Heo if (is_cmd) 3469977126cSTejun Heo fis[1] |= (1 << 7); /* bit 7 indicates Command FIS */ 3479977126cSTejun Heo 348c6fd2807SJeff Garzik fis[2] = tf->command; 349c6fd2807SJeff Garzik fis[3] = tf->feature; 350c6fd2807SJeff Garzik 351c6fd2807SJeff Garzik fis[4] = tf->lbal; 352c6fd2807SJeff Garzik fis[5] = tf->lbam; 353c6fd2807SJeff Garzik fis[6] = tf->lbah; 354c6fd2807SJeff Garzik fis[7] = tf->device; 355c6fd2807SJeff Garzik 356c6fd2807SJeff Garzik fis[8] = tf->hob_lbal; 357c6fd2807SJeff Garzik fis[9] = tf->hob_lbam; 358c6fd2807SJeff Garzik fis[10] = tf->hob_lbah; 359c6fd2807SJeff Garzik fis[11] = tf->hob_feature; 360c6fd2807SJeff Garzik 361c6fd2807SJeff Garzik fis[12] = tf->nsect; 362c6fd2807SJeff Garzik fis[13] = tf->hob_nsect; 363c6fd2807SJeff Garzik fis[14] = 0; 364c6fd2807SJeff Garzik fis[15] = tf->ctl; 365c6fd2807SJeff Garzik 366c6fd2807SJeff Garzik fis[16] = 0; 367c6fd2807SJeff Garzik fis[17] = 0; 368c6fd2807SJeff Garzik fis[18] = 0; 369c6fd2807SJeff Garzik fis[19] = 0; 370c6fd2807SJeff Garzik } 371c6fd2807SJeff Garzik 372c6fd2807SJeff Garzik /** 373c6fd2807SJeff Garzik * ata_tf_from_fis - Convert SATA FIS to ATA taskfile 374c6fd2807SJeff Garzik * @fis: Buffer from which data will be input 375c6fd2807SJeff Garzik * @tf: Taskfile to output 376c6fd2807SJeff Garzik * 377c6fd2807SJeff Garzik * Converts a serial ATA FIS structure to a standard ATA taskfile. 378c6fd2807SJeff Garzik * 379c6fd2807SJeff Garzik * LOCKING: 380c6fd2807SJeff Garzik * Inherited from caller. 381c6fd2807SJeff Garzik */ 382c6fd2807SJeff Garzik 383c6fd2807SJeff Garzik void ata_tf_from_fis(const u8 *fis, struct ata_taskfile *tf) 384c6fd2807SJeff Garzik { 385c6fd2807SJeff Garzik tf->command = fis[2]; /* status */ 386c6fd2807SJeff Garzik tf->feature = fis[3]; /* error */ 387c6fd2807SJeff Garzik 388c6fd2807SJeff Garzik tf->lbal = fis[4]; 389c6fd2807SJeff Garzik tf->lbam = fis[5]; 390c6fd2807SJeff Garzik tf->lbah = fis[6]; 391c6fd2807SJeff Garzik tf->device = fis[7]; 392c6fd2807SJeff Garzik 393c6fd2807SJeff Garzik tf->hob_lbal = fis[8]; 394c6fd2807SJeff Garzik tf->hob_lbam = fis[9]; 395c6fd2807SJeff Garzik tf->hob_lbah = fis[10]; 396c6fd2807SJeff Garzik 397c6fd2807SJeff Garzik tf->nsect = fis[12]; 398c6fd2807SJeff Garzik tf->hob_nsect = fis[13]; 399c6fd2807SJeff Garzik } 400c6fd2807SJeff Garzik 401c6fd2807SJeff Garzik static const u8 ata_rw_cmds[] = { 402c6fd2807SJeff Garzik /* pio multi */ 403c6fd2807SJeff Garzik ATA_CMD_READ_MULTI, 404c6fd2807SJeff Garzik ATA_CMD_WRITE_MULTI, 405c6fd2807SJeff Garzik ATA_CMD_READ_MULTI_EXT, 406c6fd2807SJeff Garzik ATA_CMD_WRITE_MULTI_EXT, 407c6fd2807SJeff Garzik 0, 408c6fd2807SJeff Garzik 0, 409c6fd2807SJeff Garzik 0, 410c6fd2807SJeff Garzik ATA_CMD_WRITE_MULTI_FUA_EXT, 411c6fd2807SJeff Garzik /* pio */ 412c6fd2807SJeff Garzik ATA_CMD_PIO_READ, 413c6fd2807SJeff Garzik ATA_CMD_PIO_WRITE, 414c6fd2807SJeff Garzik ATA_CMD_PIO_READ_EXT, 415c6fd2807SJeff Garzik ATA_CMD_PIO_WRITE_EXT, 416c6fd2807SJeff Garzik 0, 417c6fd2807SJeff Garzik 0, 418c6fd2807SJeff Garzik 0, 419c6fd2807SJeff Garzik 0, 420c6fd2807SJeff Garzik /* dma */ 421c6fd2807SJeff Garzik ATA_CMD_READ, 422c6fd2807SJeff Garzik ATA_CMD_WRITE, 423c6fd2807SJeff Garzik ATA_CMD_READ_EXT, 424c6fd2807SJeff Garzik ATA_CMD_WRITE_EXT, 425c6fd2807SJeff Garzik 0, 426c6fd2807SJeff Garzik 0, 427c6fd2807SJeff Garzik 0, 428c6fd2807SJeff Garzik ATA_CMD_WRITE_FUA_EXT 429c6fd2807SJeff Garzik }; 430c6fd2807SJeff Garzik 431c6fd2807SJeff Garzik /** 432c6fd2807SJeff Garzik * ata_rwcmd_protocol - set taskfile r/w commands and protocol 433bd056d7eSTejun Heo * @tf: command to examine and configure 434bd056d7eSTejun Heo * @dev: device tf belongs to 435c6fd2807SJeff Garzik * 436c6fd2807SJeff Garzik * Examine the device configuration and tf->flags to calculate 437c6fd2807SJeff Garzik * the proper read/write commands and protocol to use. 438c6fd2807SJeff Garzik * 439c6fd2807SJeff Garzik * LOCKING: 440c6fd2807SJeff Garzik * caller. 441c6fd2807SJeff Garzik */ 442bd056d7eSTejun Heo static int ata_rwcmd_protocol(struct ata_taskfile *tf, struct ata_device *dev) 443c6fd2807SJeff Garzik { 444c6fd2807SJeff Garzik u8 cmd; 445c6fd2807SJeff Garzik 446c6fd2807SJeff Garzik int index, fua, lba48, write; 447c6fd2807SJeff Garzik 448c6fd2807SJeff Garzik fua = (tf->flags & ATA_TFLAG_FUA) ? 4 : 0; 449c6fd2807SJeff Garzik lba48 = (tf->flags & ATA_TFLAG_LBA48) ? 2 : 0; 450c6fd2807SJeff Garzik write = (tf->flags & ATA_TFLAG_WRITE) ? 1 : 0; 451c6fd2807SJeff Garzik 452c6fd2807SJeff Garzik if (dev->flags & ATA_DFLAG_PIO) { 453c6fd2807SJeff Garzik tf->protocol = ATA_PROT_PIO; 454c6fd2807SJeff Garzik index = dev->multi_count ? 0 : 8; 4559af5c9c9STejun Heo } else if (lba48 && (dev->link->ap->flags & ATA_FLAG_PIO_LBA48)) { 456c6fd2807SJeff Garzik /* Unable to use DMA due to host limitation */ 457c6fd2807SJeff Garzik tf->protocol = ATA_PROT_PIO; 458c6fd2807SJeff Garzik index = dev->multi_count ? 0 : 8; 459c6fd2807SJeff Garzik } else { 460c6fd2807SJeff Garzik tf->protocol = ATA_PROT_DMA; 461c6fd2807SJeff Garzik index = 16; 462c6fd2807SJeff Garzik } 463c6fd2807SJeff Garzik 464c6fd2807SJeff Garzik cmd = ata_rw_cmds[index + fua + lba48 + write]; 465c6fd2807SJeff Garzik if (cmd) { 466c6fd2807SJeff Garzik tf->command = cmd; 467c6fd2807SJeff Garzik return 0; 468c6fd2807SJeff Garzik } 469c6fd2807SJeff Garzik return -1; 470c6fd2807SJeff Garzik } 471c6fd2807SJeff Garzik 472c6fd2807SJeff Garzik /** 47335b649feSTejun Heo * ata_tf_read_block - Read block address from ATA taskfile 47435b649feSTejun Heo * @tf: ATA taskfile of interest 47535b649feSTejun Heo * @dev: ATA device @tf belongs to 47635b649feSTejun Heo * 47735b649feSTejun Heo * LOCKING: 47835b649feSTejun Heo * None. 47935b649feSTejun Heo * 48035b649feSTejun Heo * Read block address from @tf. This function can handle all 48135b649feSTejun Heo * three address formats - LBA, LBA48 and CHS. tf->protocol and 48235b649feSTejun Heo * flags select the address format to use. 48335b649feSTejun Heo * 48435b649feSTejun Heo * RETURNS: 48535b649feSTejun Heo * Block address read from @tf. 48635b649feSTejun Heo */ 48735b649feSTejun Heo u64 ata_tf_read_block(struct ata_taskfile *tf, struct ata_device *dev) 48835b649feSTejun Heo { 48935b649feSTejun Heo u64 block = 0; 49035b649feSTejun Heo 49135b649feSTejun Heo if (tf->flags & ATA_TFLAG_LBA) { 49235b649feSTejun Heo if (tf->flags & ATA_TFLAG_LBA48) { 49335b649feSTejun Heo block |= (u64)tf->hob_lbah << 40; 49435b649feSTejun Heo block |= (u64)tf->hob_lbam << 32; 49535b649feSTejun Heo block |= tf->hob_lbal << 24; 49635b649feSTejun Heo } else 49735b649feSTejun Heo block |= (tf->device & 0xf) << 24; 49835b649feSTejun Heo 49935b649feSTejun Heo block |= tf->lbah << 16; 50035b649feSTejun Heo block |= tf->lbam << 8; 50135b649feSTejun Heo block |= tf->lbal; 50235b649feSTejun Heo } else { 50335b649feSTejun Heo u32 cyl, head, sect; 50435b649feSTejun Heo 50535b649feSTejun Heo cyl = tf->lbam | (tf->lbah << 8); 50635b649feSTejun Heo head = tf->device & 0xf; 50735b649feSTejun Heo sect = tf->lbal; 50835b649feSTejun Heo 50935b649feSTejun Heo block = (cyl * dev->heads + head) * dev->sectors + sect; 51035b649feSTejun Heo } 51135b649feSTejun Heo 51235b649feSTejun Heo return block; 51335b649feSTejun Heo } 51435b649feSTejun Heo 51535b649feSTejun Heo /** 516bd056d7eSTejun Heo * ata_build_rw_tf - Build ATA taskfile for given read/write request 517bd056d7eSTejun Heo * @tf: Target ATA taskfile 518bd056d7eSTejun Heo * @dev: ATA device @tf belongs to 519bd056d7eSTejun Heo * @block: Block address 520bd056d7eSTejun Heo * @n_block: Number of blocks 521bd056d7eSTejun Heo * @tf_flags: RW/FUA etc... 522bd056d7eSTejun Heo * @tag: tag 523bd056d7eSTejun Heo * 524bd056d7eSTejun Heo * LOCKING: 525bd056d7eSTejun Heo * None. 526bd056d7eSTejun Heo * 527bd056d7eSTejun Heo * Build ATA taskfile @tf for read/write request described by 528bd056d7eSTejun Heo * @block, @n_block, @tf_flags and @tag on @dev. 529bd056d7eSTejun Heo * 530bd056d7eSTejun Heo * RETURNS: 531bd056d7eSTejun Heo * 532bd056d7eSTejun Heo * 0 on success, -ERANGE if the request is too large for @dev, 533bd056d7eSTejun Heo * -EINVAL if the request is invalid. 534bd056d7eSTejun Heo */ 535bd056d7eSTejun Heo int ata_build_rw_tf(struct ata_taskfile *tf, struct ata_device *dev, 536bd056d7eSTejun Heo u64 block, u32 n_block, unsigned int tf_flags, 537bd056d7eSTejun Heo unsigned int tag) 538bd056d7eSTejun Heo { 539bd056d7eSTejun Heo tf->flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; 540bd056d7eSTejun Heo tf->flags |= tf_flags; 541bd056d7eSTejun Heo 5426d1245bfSTejun Heo if (ata_ncq_enabled(dev) && likely(tag != ATA_TAG_INTERNAL)) { 543bd056d7eSTejun Heo /* yay, NCQ */ 544bd056d7eSTejun Heo if (!lba_48_ok(block, n_block)) 545bd056d7eSTejun Heo return -ERANGE; 546bd056d7eSTejun Heo 547bd056d7eSTejun Heo tf->protocol = ATA_PROT_NCQ; 548bd056d7eSTejun Heo tf->flags |= ATA_TFLAG_LBA | ATA_TFLAG_LBA48; 549bd056d7eSTejun Heo 550bd056d7eSTejun Heo if (tf->flags & ATA_TFLAG_WRITE) 551bd056d7eSTejun Heo tf->command = ATA_CMD_FPDMA_WRITE; 552bd056d7eSTejun Heo else 553bd056d7eSTejun Heo tf->command = ATA_CMD_FPDMA_READ; 554bd056d7eSTejun Heo 555bd056d7eSTejun Heo tf->nsect = tag << 3; 556bd056d7eSTejun Heo tf->hob_feature = (n_block >> 8) & 0xff; 557bd056d7eSTejun Heo tf->feature = n_block & 0xff; 558bd056d7eSTejun Heo 559bd056d7eSTejun Heo tf->hob_lbah = (block >> 40) & 0xff; 560bd056d7eSTejun Heo tf->hob_lbam = (block >> 32) & 0xff; 561bd056d7eSTejun Heo tf->hob_lbal = (block >> 24) & 0xff; 562bd056d7eSTejun Heo tf->lbah = (block >> 16) & 0xff; 563bd056d7eSTejun Heo tf->lbam = (block >> 8) & 0xff; 564bd056d7eSTejun Heo tf->lbal = block & 0xff; 565bd056d7eSTejun Heo 566bd056d7eSTejun Heo tf->device = 1 << 6; 567bd056d7eSTejun Heo if (tf->flags & ATA_TFLAG_FUA) 568bd056d7eSTejun Heo tf->device |= 1 << 7; 569bd056d7eSTejun Heo } else if (dev->flags & ATA_DFLAG_LBA) { 570bd056d7eSTejun Heo tf->flags |= ATA_TFLAG_LBA; 571bd056d7eSTejun Heo 572bd056d7eSTejun Heo if (lba_28_ok(block, n_block)) { 573bd056d7eSTejun Heo /* use LBA28 */ 574bd056d7eSTejun Heo tf->device |= (block >> 24) & 0xf; 575bd056d7eSTejun Heo } else if (lba_48_ok(block, n_block)) { 576bd056d7eSTejun Heo if (!(dev->flags & ATA_DFLAG_LBA48)) 577bd056d7eSTejun Heo return -ERANGE; 578bd056d7eSTejun Heo 579bd056d7eSTejun Heo /* use LBA48 */ 580bd056d7eSTejun Heo tf->flags |= ATA_TFLAG_LBA48; 581bd056d7eSTejun Heo 582bd056d7eSTejun Heo tf->hob_nsect = (n_block >> 8) & 0xff; 583bd056d7eSTejun Heo 584bd056d7eSTejun Heo tf->hob_lbah = (block >> 40) & 0xff; 585bd056d7eSTejun Heo tf->hob_lbam = (block >> 32) & 0xff; 586bd056d7eSTejun Heo tf->hob_lbal = (block >> 24) & 0xff; 587bd056d7eSTejun Heo } else 588bd056d7eSTejun Heo /* request too large even for LBA48 */ 589bd056d7eSTejun Heo return -ERANGE; 590bd056d7eSTejun Heo 591bd056d7eSTejun Heo if (unlikely(ata_rwcmd_protocol(tf, dev) < 0)) 592bd056d7eSTejun Heo return -EINVAL; 593bd056d7eSTejun Heo 594bd056d7eSTejun Heo tf->nsect = n_block & 0xff; 595bd056d7eSTejun Heo 596bd056d7eSTejun Heo tf->lbah = (block >> 16) & 0xff; 597bd056d7eSTejun Heo tf->lbam = (block >> 8) & 0xff; 598bd056d7eSTejun Heo tf->lbal = block & 0xff; 599bd056d7eSTejun Heo 600bd056d7eSTejun Heo tf->device |= ATA_LBA; 601bd056d7eSTejun Heo } else { 602bd056d7eSTejun Heo /* CHS */ 603bd056d7eSTejun Heo u32 sect, head, cyl, track; 604bd056d7eSTejun Heo 605bd056d7eSTejun Heo /* The request -may- be too large for CHS addressing. */ 606bd056d7eSTejun Heo if (!lba_28_ok(block, n_block)) 607bd056d7eSTejun Heo return -ERANGE; 608bd056d7eSTejun Heo 609bd056d7eSTejun Heo if (unlikely(ata_rwcmd_protocol(tf, dev) < 0)) 610bd056d7eSTejun Heo return -EINVAL; 611bd056d7eSTejun Heo 612bd056d7eSTejun Heo /* Convert LBA to CHS */ 613bd056d7eSTejun Heo track = (u32)block / dev->sectors; 614bd056d7eSTejun Heo cyl = track / dev->heads; 615bd056d7eSTejun Heo head = track % dev->heads; 616bd056d7eSTejun Heo sect = (u32)block % dev->sectors + 1; 617bd056d7eSTejun Heo 618bd056d7eSTejun Heo DPRINTK("block %u track %u cyl %u head %u sect %u\n", 619bd056d7eSTejun Heo (u32)block, track, cyl, head, sect); 620bd056d7eSTejun Heo 621bd056d7eSTejun Heo /* Check whether the converted CHS can fit. 622bd056d7eSTejun Heo Cylinder: 0-65535 623bd056d7eSTejun Heo Head: 0-15 624bd056d7eSTejun Heo Sector: 1-255*/ 625bd056d7eSTejun Heo if ((cyl >> 16) || (head >> 4) || (sect >> 8) || (!sect)) 626bd056d7eSTejun Heo return -ERANGE; 627bd056d7eSTejun Heo 628bd056d7eSTejun Heo tf->nsect = n_block & 0xff; /* Sector count 0 means 256 sectors */ 629bd056d7eSTejun Heo tf->lbal = sect; 630bd056d7eSTejun Heo tf->lbam = cyl; 631bd056d7eSTejun Heo tf->lbah = cyl >> 8; 632bd056d7eSTejun Heo tf->device |= head; 633bd056d7eSTejun Heo } 634bd056d7eSTejun Heo 635bd056d7eSTejun Heo return 0; 636bd056d7eSTejun Heo } 637bd056d7eSTejun Heo 638bd056d7eSTejun Heo /** 639c6fd2807SJeff Garzik * ata_pack_xfermask - Pack pio, mwdma and udma masks into xfer_mask 640c6fd2807SJeff Garzik * @pio_mask: pio_mask 641c6fd2807SJeff Garzik * @mwdma_mask: mwdma_mask 642c6fd2807SJeff Garzik * @udma_mask: udma_mask 643c6fd2807SJeff Garzik * 644c6fd2807SJeff Garzik * Pack @pio_mask, @mwdma_mask and @udma_mask into a single 645c6fd2807SJeff Garzik * unsigned int xfer_mask. 646c6fd2807SJeff Garzik * 647c6fd2807SJeff Garzik * LOCKING: 648c6fd2807SJeff Garzik * None. 649c6fd2807SJeff Garzik * 650c6fd2807SJeff Garzik * RETURNS: 651c6fd2807SJeff Garzik * Packed xfer_mask. 652c6fd2807SJeff Garzik */ 6537dc951aeSTejun Heo unsigned long ata_pack_xfermask(unsigned long pio_mask, 6547dc951aeSTejun Heo unsigned long mwdma_mask, 6557dc951aeSTejun Heo unsigned long udma_mask) 656c6fd2807SJeff Garzik { 657c6fd2807SJeff Garzik return ((pio_mask << ATA_SHIFT_PIO) & ATA_MASK_PIO) | 658c6fd2807SJeff Garzik ((mwdma_mask << ATA_SHIFT_MWDMA) & ATA_MASK_MWDMA) | 659c6fd2807SJeff Garzik ((udma_mask << ATA_SHIFT_UDMA) & ATA_MASK_UDMA); 660c6fd2807SJeff Garzik } 661c6fd2807SJeff Garzik 662c6fd2807SJeff Garzik /** 663c6fd2807SJeff Garzik * ata_unpack_xfermask - Unpack xfer_mask into pio, mwdma and udma masks 664c6fd2807SJeff Garzik * @xfer_mask: xfer_mask to unpack 665c6fd2807SJeff Garzik * @pio_mask: resulting pio_mask 666c6fd2807SJeff Garzik * @mwdma_mask: resulting mwdma_mask 667c6fd2807SJeff Garzik * @udma_mask: resulting udma_mask 668c6fd2807SJeff Garzik * 669c6fd2807SJeff Garzik * Unpack @xfer_mask into @pio_mask, @mwdma_mask and @udma_mask. 670c6fd2807SJeff Garzik * Any NULL distination masks will be ignored. 671c6fd2807SJeff Garzik */ 6727dc951aeSTejun Heo void ata_unpack_xfermask(unsigned long xfer_mask, unsigned long *pio_mask, 6737dc951aeSTejun Heo unsigned long *mwdma_mask, unsigned long *udma_mask) 674c6fd2807SJeff Garzik { 675c6fd2807SJeff Garzik if (pio_mask) 676c6fd2807SJeff Garzik *pio_mask = (xfer_mask & ATA_MASK_PIO) >> ATA_SHIFT_PIO; 677c6fd2807SJeff Garzik if (mwdma_mask) 678c6fd2807SJeff Garzik *mwdma_mask = (xfer_mask & ATA_MASK_MWDMA) >> ATA_SHIFT_MWDMA; 679c6fd2807SJeff Garzik if (udma_mask) 680c6fd2807SJeff Garzik *udma_mask = (xfer_mask & ATA_MASK_UDMA) >> ATA_SHIFT_UDMA; 681c6fd2807SJeff Garzik } 682c6fd2807SJeff Garzik 683c6fd2807SJeff Garzik static const struct ata_xfer_ent { 684c6fd2807SJeff Garzik int shift, bits; 685c6fd2807SJeff Garzik u8 base; 686c6fd2807SJeff Garzik } ata_xfer_tbl[] = { 68770cd071eSTejun Heo { ATA_SHIFT_PIO, ATA_NR_PIO_MODES, XFER_PIO_0 }, 68870cd071eSTejun Heo { ATA_SHIFT_MWDMA, ATA_NR_MWDMA_MODES, XFER_MW_DMA_0 }, 68970cd071eSTejun Heo { ATA_SHIFT_UDMA, ATA_NR_UDMA_MODES, XFER_UDMA_0 }, 690c6fd2807SJeff Garzik { -1, }, 691c6fd2807SJeff Garzik }; 692c6fd2807SJeff Garzik 693c6fd2807SJeff Garzik /** 694c6fd2807SJeff Garzik * ata_xfer_mask2mode - Find matching XFER_* for the given xfer_mask 695c6fd2807SJeff Garzik * @xfer_mask: xfer_mask of interest 696c6fd2807SJeff Garzik * 697c6fd2807SJeff Garzik * Return matching XFER_* value for @xfer_mask. Only the highest 698c6fd2807SJeff Garzik * bit of @xfer_mask is considered. 699c6fd2807SJeff Garzik * 700c6fd2807SJeff Garzik * LOCKING: 701c6fd2807SJeff Garzik * None. 702c6fd2807SJeff Garzik * 703c6fd2807SJeff Garzik * RETURNS: 70470cd071eSTejun Heo * Matching XFER_* value, 0xff if no match found. 705c6fd2807SJeff Garzik */ 7067dc951aeSTejun Heo u8 ata_xfer_mask2mode(unsigned long xfer_mask) 707c6fd2807SJeff Garzik { 708c6fd2807SJeff Garzik int highbit = fls(xfer_mask) - 1; 709c6fd2807SJeff Garzik const struct ata_xfer_ent *ent; 710c6fd2807SJeff Garzik 711c6fd2807SJeff Garzik for (ent = ata_xfer_tbl; ent->shift >= 0; ent++) 712c6fd2807SJeff Garzik if (highbit >= ent->shift && highbit < ent->shift + ent->bits) 713c6fd2807SJeff Garzik return ent->base + highbit - ent->shift; 71470cd071eSTejun Heo return 0xff; 715c6fd2807SJeff Garzik } 716c6fd2807SJeff Garzik 717c6fd2807SJeff Garzik /** 718c6fd2807SJeff Garzik * ata_xfer_mode2mask - Find matching xfer_mask for XFER_* 719c6fd2807SJeff Garzik * @xfer_mode: XFER_* of interest 720c6fd2807SJeff Garzik * 721c6fd2807SJeff Garzik * Return matching xfer_mask for @xfer_mode. 722c6fd2807SJeff Garzik * 723c6fd2807SJeff Garzik * LOCKING: 724c6fd2807SJeff Garzik * None. 725c6fd2807SJeff Garzik * 726c6fd2807SJeff Garzik * RETURNS: 727c6fd2807SJeff Garzik * Matching xfer_mask, 0 if no match found. 728c6fd2807SJeff Garzik */ 7297dc951aeSTejun Heo unsigned long ata_xfer_mode2mask(u8 xfer_mode) 730c6fd2807SJeff Garzik { 731c6fd2807SJeff Garzik const struct ata_xfer_ent *ent; 732c6fd2807SJeff Garzik 733c6fd2807SJeff Garzik for (ent = ata_xfer_tbl; ent->shift >= 0; ent++) 734c6fd2807SJeff Garzik if (xfer_mode >= ent->base && xfer_mode < ent->base + ent->bits) 73570cd071eSTejun Heo return ((2 << (ent->shift + xfer_mode - ent->base)) - 1) 73670cd071eSTejun Heo & ~((1 << ent->shift) - 1); 737c6fd2807SJeff Garzik return 0; 738c6fd2807SJeff Garzik } 739c6fd2807SJeff Garzik 740c6fd2807SJeff Garzik /** 741c6fd2807SJeff Garzik * ata_xfer_mode2shift - Find matching xfer_shift for XFER_* 742c6fd2807SJeff Garzik * @xfer_mode: XFER_* of interest 743c6fd2807SJeff Garzik * 744c6fd2807SJeff Garzik * Return matching xfer_shift for @xfer_mode. 745c6fd2807SJeff Garzik * 746c6fd2807SJeff Garzik * LOCKING: 747c6fd2807SJeff Garzik * None. 748c6fd2807SJeff Garzik * 749c6fd2807SJeff Garzik * RETURNS: 750c6fd2807SJeff Garzik * Matching xfer_shift, -1 if no match found. 751c6fd2807SJeff Garzik */ 7527dc951aeSTejun Heo int ata_xfer_mode2shift(unsigned long xfer_mode) 753c6fd2807SJeff Garzik { 754c6fd2807SJeff Garzik const struct ata_xfer_ent *ent; 755c6fd2807SJeff Garzik 756c6fd2807SJeff Garzik for (ent = ata_xfer_tbl; ent->shift >= 0; ent++) 757c6fd2807SJeff Garzik if (xfer_mode >= ent->base && xfer_mode < ent->base + ent->bits) 758c6fd2807SJeff Garzik return ent->shift; 759c6fd2807SJeff Garzik return -1; 760c6fd2807SJeff Garzik } 761c6fd2807SJeff Garzik 762c6fd2807SJeff Garzik /** 763c6fd2807SJeff Garzik * ata_mode_string - convert xfer_mask to string 764c6fd2807SJeff Garzik * @xfer_mask: mask of bits supported; only highest bit counts. 765c6fd2807SJeff Garzik * 766c6fd2807SJeff Garzik * Determine string which represents the highest speed 767c6fd2807SJeff Garzik * (highest bit in @modemask). 768c6fd2807SJeff Garzik * 769c6fd2807SJeff Garzik * LOCKING: 770c6fd2807SJeff Garzik * None. 771c6fd2807SJeff Garzik * 772c6fd2807SJeff Garzik * RETURNS: 773c6fd2807SJeff Garzik * Constant C string representing highest speed listed in 774c6fd2807SJeff Garzik * @mode_mask, or the constant C string "<n/a>". 775c6fd2807SJeff Garzik */ 7767dc951aeSTejun Heo const char *ata_mode_string(unsigned long xfer_mask) 777c6fd2807SJeff Garzik { 778c6fd2807SJeff Garzik static const char * const xfer_mode_str[] = { 779c6fd2807SJeff Garzik "PIO0", 780c6fd2807SJeff Garzik "PIO1", 781c6fd2807SJeff Garzik "PIO2", 782c6fd2807SJeff Garzik "PIO3", 783c6fd2807SJeff Garzik "PIO4", 784b352e57dSAlan Cox "PIO5", 785b352e57dSAlan Cox "PIO6", 786c6fd2807SJeff Garzik "MWDMA0", 787c6fd2807SJeff Garzik "MWDMA1", 788c6fd2807SJeff Garzik "MWDMA2", 789b352e57dSAlan Cox "MWDMA3", 790b352e57dSAlan Cox "MWDMA4", 791c6fd2807SJeff Garzik "UDMA/16", 792c6fd2807SJeff Garzik "UDMA/25", 793c6fd2807SJeff Garzik "UDMA/33", 794c6fd2807SJeff Garzik "UDMA/44", 795c6fd2807SJeff Garzik "UDMA/66", 796c6fd2807SJeff Garzik "UDMA/100", 797c6fd2807SJeff Garzik "UDMA/133", 798c6fd2807SJeff Garzik "UDMA7", 799c6fd2807SJeff Garzik }; 800c6fd2807SJeff Garzik int highbit; 801c6fd2807SJeff Garzik 802c6fd2807SJeff Garzik highbit = fls(xfer_mask) - 1; 803c6fd2807SJeff Garzik if (highbit >= 0 && highbit < ARRAY_SIZE(xfer_mode_str)) 804c6fd2807SJeff Garzik return xfer_mode_str[highbit]; 805c6fd2807SJeff Garzik return "<n/a>"; 806c6fd2807SJeff Garzik } 807c6fd2807SJeff Garzik 808c6fd2807SJeff Garzik static const char *sata_spd_string(unsigned int spd) 809c6fd2807SJeff Garzik { 810c6fd2807SJeff Garzik static const char * const spd_str[] = { 811c6fd2807SJeff Garzik "1.5 Gbps", 812c6fd2807SJeff Garzik "3.0 Gbps", 813c6fd2807SJeff Garzik }; 814c6fd2807SJeff Garzik 815c6fd2807SJeff Garzik if (spd == 0 || (spd - 1) >= ARRAY_SIZE(spd_str)) 816c6fd2807SJeff Garzik return "<unknown>"; 817c6fd2807SJeff Garzik return spd_str[spd - 1]; 818c6fd2807SJeff Garzik } 819c6fd2807SJeff Garzik 820c6fd2807SJeff Garzik void ata_dev_disable(struct ata_device *dev) 821c6fd2807SJeff Garzik { 82209d7f9b0STejun Heo if (ata_dev_enabled(dev)) { 8239af5c9c9STejun Heo if (ata_msg_drv(dev->link->ap)) 824c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_WARNING, "disabled\n"); 825562f0c2dSTejun Heo ata_acpi_on_disable(dev); 8264ae72a1eSTejun Heo ata_down_xfermask_limit(dev, ATA_DNXFER_FORCE_PIO0 | 8274ae72a1eSTejun Heo ATA_DNXFER_QUIET); 828c6fd2807SJeff Garzik dev->class++; 829c6fd2807SJeff Garzik } 830c6fd2807SJeff Garzik } 831c6fd2807SJeff Garzik 832ca77329fSKristen Carlson Accardi static int ata_dev_set_dipm(struct ata_device *dev, enum link_pm policy) 833ca77329fSKristen Carlson Accardi { 834ca77329fSKristen Carlson Accardi struct ata_link *link = dev->link; 835ca77329fSKristen Carlson Accardi struct ata_port *ap = link->ap; 836ca77329fSKristen Carlson Accardi u32 scontrol; 837ca77329fSKristen Carlson Accardi unsigned int err_mask; 838ca77329fSKristen Carlson Accardi int rc; 839ca77329fSKristen Carlson Accardi 840ca77329fSKristen Carlson Accardi /* 841ca77329fSKristen Carlson Accardi * disallow DIPM for drivers which haven't set 842ca77329fSKristen Carlson Accardi * ATA_FLAG_IPM. This is because when DIPM is enabled, 843ca77329fSKristen Carlson Accardi * phy ready will be set in the interrupt status on 844ca77329fSKristen Carlson Accardi * state changes, which will cause some drivers to 845ca77329fSKristen Carlson Accardi * think there are errors - additionally drivers will 846ca77329fSKristen Carlson Accardi * need to disable hot plug. 847ca77329fSKristen Carlson Accardi */ 848ca77329fSKristen Carlson Accardi if (!(ap->flags & ATA_FLAG_IPM) || !ata_dev_enabled(dev)) { 849ca77329fSKristen Carlson Accardi ap->pm_policy = NOT_AVAILABLE; 850ca77329fSKristen Carlson Accardi return -EINVAL; 851ca77329fSKristen Carlson Accardi } 852ca77329fSKristen Carlson Accardi 853ca77329fSKristen Carlson Accardi /* 854ca77329fSKristen Carlson Accardi * For DIPM, we will only enable it for the 855ca77329fSKristen Carlson Accardi * min_power setting. 856ca77329fSKristen Carlson Accardi * 857ca77329fSKristen Carlson Accardi * Why? Because Disks are too stupid to know that 858ca77329fSKristen Carlson Accardi * If the host rejects a request to go to SLUMBER 859ca77329fSKristen Carlson Accardi * they should retry at PARTIAL, and instead it 860ca77329fSKristen Carlson Accardi * just would give up. So, for medium_power to 861ca77329fSKristen Carlson Accardi * work at all, we need to only allow HIPM. 862ca77329fSKristen Carlson Accardi */ 863ca77329fSKristen Carlson Accardi rc = sata_scr_read(link, SCR_CONTROL, &scontrol); 864ca77329fSKristen Carlson Accardi if (rc) 865ca77329fSKristen Carlson Accardi return rc; 866ca77329fSKristen Carlson Accardi 867ca77329fSKristen Carlson Accardi switch (policy) { 868ca77329fSKristen Carlson Accardi case MIN_POWER: 869ca77329fSKristen Carlson Accardi /* no restrictions on IPM transitions */ 870ca77329fSKristen Carlson Accardi scontrol &= ~(0x3 << 8); 871ca77329fSKristen Carlson Accardi rc = sata_scr_write(link, SCR_CONTROL, scontrol); 872ca77329fSKristen Carlson Accardi if (rc) 873ca77329fSKristen Carlson Accardi return rc; 874ca77329fSKristen Carlson Accardi 875ca77329fSKristen Carlson Accardi /* enable DIPM */ 876ca77329fSKristen Carlson Accardi if (dev->flags & ATA_DFLAG_DIPM) 877ca77329fSKristen Carlson Accardi err_mask = ata_dev_set_feature(dev, 878ca77329fSKristen Carlson Accardi SETFEATURES_SATA_ENABLE, SATA_DIPM); 879ca77329fSKristen Carlson Accardi break; 880ca77329fSKristen Carlson Accardi case MEDIUM_POWER: 881ca77329fSKristen Carlson Accardi /* allow IPM to PARTIAL */ 882ca77329fSKristen Carlson Accardi scontrol &= ~(0x1 << 8); 883ca77329fSKristen Carlson Accardi scontrol |= (0x2 << 8); 884ca77329fSKristen Carlson Accardi rc = sata_scr_write(link, SCR_CONTROL, scontrol); 885ca77329fSKristen Carlson Accardi if (rc) 886ca77329fSKristen Carlson Accardi return rc; 887ca77329fSKristen Carlson Accardi 888f5456b63SKristen Carlson Accardi /* 889f5456b63SKristen Carlson Accardi * we don't have to disable DIPM since IPM flags 890f5456b63SKristen Carlson Accardi * disallow transitions to SLUMBER, which effectively 891f5456b63SKristen Carlson Accardi * disable DIPM if it does not support PARTIAL 892f5456b63SKristen Carlson Accardi */ 893ca77329fSKristen Carlson Accardi break; 894ca77329fSKristen Carlson Accardi case NOT_AVAILABLE: 895ca77329fSKristen Carlson Accardi case MAX_PERFORMANCE: 896ca77329fSKristen Carlson Accardi /* disable all IPM transitions */ 897ca77329fSKristen Carlson Accardi scontrol |= (0x3 << 8); 898ca77329fSKristen Carlson Accardi rc = sata_scr_write(link, SCR_CONTROL, scontrol); 899ca77329fSKristen Carlson Accardi if (rc) 900ca77329fSKristen Carlson Accardi return rc; 901ca77329fSKristen Carlson Accardi 902f5456b63SKristen Carlson Accardi /* 903f5456b63SKristen Carlson Accardi * we don't have to disable DIPM since IPM flags 904f5456b63SKristen Carlson Accardi * disallow all transitions which effectively 905f5456b63SKristen Carlson Accardi * disable DIPM anyway. 906f5456b63SKristen Carlson Accardi */ 907ca77329fSKristen Carlson Accardi break; 908ca77329fSKristen Carlson Accardi } 909ca77329fSKristen Carlson Accardi 910ca77329fSKristen Carlson Accardi /* FIXME: handle SET FEATURES failure */ 911ca77329fSKristen Carlson Accardi (void) err_mask; 912ca77329fSKristen Carlson Accardi 913ca77329fSKristen Carlson Accardi return 0; 914ca77329fSKristen Carlson Accardi } 915ca77329fSKristen Carlson Accardi 916ca77329fSKristen Carlson Accardi /** 917ca77329fSKristen Carlson Accardi * ata_dev_enable_pm - enable SATA interface power management 91848166fd9SStephen Hemminger * @dev: device to enable power management 91948166fd9SStephen Hemminger * @policy: the link power management policy 920ca77329fSKristen Carlson Accardi * 921ca77329fSKristen Carlson Accardi * Enable SATA Interface power management. This will enable 922ca77329fSKristen Carlson Accardi * Device Interface Power Management (DIPM) for min_power 923ca77329fSKristen Carlson Accardi * policy, and then call driver specific callbacks for 924ca77329fSKristen Carlson Accardi * enabling Host Initiated Power management. 925ca77329fSKristen Carlson Accardi * 926ca77329fSKristen Carlson Accardi * Locking: Caller. 927ca77329fSKristen Carlson Accardi * Returns: -EINVAL if IPM is not supported, 0 otherwise. 928ca77329fSKristen Carlson Accardi */ 929ca77329fSKristen Carlson Accardi void ata_dev_enable_pm(struct ata_device *dev, enum link_pm policy) 930ca77329fSKristen Carlson Accardi { 931ca77329fSKristen Carlson Accardi int rc = 0; 932ca77329fSKristen Carlson Accardi struct ata_port *ap = dev->link->ap; 933ca77329fSKristen Carlson Accardi 934ca77329fSKristen Carlson Accardi /* set HIPM first, then DIPM */ 935ca77329fSKristen Carlson Accardi if (ap->ops->enable_pm) 936ca77329fSKristen Carlson Accardi rc = ap->ops->enable_pm(ap, policy); 937ca77329fSKristen Carlson Accardi if (rc) 938ca77329fSKristen Carlson Accardi goto enable_pm_out; 939ca77329fSKristen Carlson Accardi rc = ata_dev_set_dipm(dev, policy); 940ca77329fSKristen Carlson Accardi 941ca77329fSKristen Carlson Accardi enable_pm_out: 942ca77329fSKristen Carlson Accardi if (rc) 943ca77329fSKristen Carlson Accardi ap->pm_policy = MAX_PERFORMANCE; 944ca77329fSKristen Carlson Accardi else 945ca77329fSKristen Carlson Accardi ap->pm_policy = policy; 946ca77329fSKristen Carlson Accardi return /* rc */; /* hopefully we can use 'rc' eventually */ 947ca77329fSKristen Carlson Accardi } 948ca77329fSKristen Carlson Accardi 9491992a5edSStephen Rothwell #ifdef CONFIG_PM 950ca77329fSKristen Carlson Accardi /** 951ca77329fSKristen Carlson Accardi * ata_dev_disable_pm - disable SATA interface power management 95248166fd9SStephen Hemminger * @dev: device to disable power management 953ca77329fSKristen Carlson Accardi * 954ca77329fSKristen Carlson Accardi * Disable SATA Interface power management. This will disable 955ca77329fSKristen Carlson Accardi * Device Interface Power Management (DIPM) without changing 956ca77329fSKristen Carlson Accardi * policy, call driver specific callbacks for disabling Host 957ca77329fSKristen Carlson Accardi * Initiated Power management. 958ca77329fSKristen Carlson Accardi * 959ca77329fSKristen Carlson Accardi * Locking: Caller. 960ca77329fSKristen Carlson Accardi * Returns: void 961ca77329fSKristen Carlson Accardi */ 962ca77329fSKristen Carlson Accardi static void ata_dev_disable_pm(struct ata_device *dev) 963ca77329fSKristen Carlson Accardi { 964ca77329fSKristen Carlson Accardi struct ata_port *ap = dev->link->ap; 965ca77329fSKristen Carlson Accardi 966ca77329fSKristen Carlson Accardi ata_dev_set_dipm(dev, MAX_PERFORMANCE); 967ca77329fSKristen Carlson Accardi if (ap->ops->disable_pm) 968ca77329fSKristen Carlson Accardi ap->ops->disable_pm(ap); 969ca77329fSKristen Carlson Accardi } 9701992a5edSStephen Rothwell #endif /* CONFIG_PM */ 971ca77329fSKristen Carlson Accardi 972ca77329fSKristen Carlson Accardi void ata_lpm_schedule(struct ata_port *ap, enum link_pm policy) 973ca77329fSKristen Carlson Accardi { 974ca77329fSKristen Carlson Accardi ap->pm_policy = policy; 975ca77329fSKristen Carlson Accardi ap->link.eh_info.action |= ATA_EHI_LPM; 976ca77329fSKristen Carlson Accardi ap->link.eh_info.flags |= ATA_EHI_NO_AUTOPSY; 977ca77329fSKristen Carlson Accardi ata_port_schedule_eh(ap); 978ca77329fSKristen Carlson Accardi } 979ca77329fSKristen Carlson Accardi 9801992a5edSStephen Rothwell #ifdef CONFIG_PM 981ca77329fSKristen Carlson Accardi static void ata_lpm_enable(struct ata_host *host) 982ca77329fSKristen Carlson Accardi { 983ca77329fSKristen Carlson Accardi struct ata_link *link; 984ca77329fSKristen Carlson Accardi struct ata_port *ap; 985ca77329fSKristen Carlson Accardi struct ata_device *dev; 986ca77329fSKristen Carlson Accardi int i; 987ca77329fSKristen Carlson Accardi 988ca77329fSKristen Carlson Accardi for (i = 0; i < host->n_ports; i++) { 989ca77329fSKristen Carlson Accardi ap = host->ports[i]; 990ca77329fSKristen Carlson Accardi ata_port_for_each_link(link, ap) { 991ca77329fSKristen Carlson Accardi ata_link_for_each_dev(dev, link) 992ca77329fSKristen Carlson Accardi ata_dev_disable_pm(dev); 993ca77329fSKristen Carlson Accardi } 994ca77329fSKristen Carlson Accardi } 995ca77329fSKristen Carlson Accardi } 996ca77329fSKristen Carlson Accardi 997ca77329fSKristen Carlson Accardi static void ata_lpm_disable(struct ata_host *host) 998ca77329fSKristen Carlson Accardi { 999ca77329fSKristen Carlson Accardi int i; 1000ca77329fSKristen Carlson Accardi 1001ca77329fSKristen Carlson Accardi for (i = 0; i < host->n_ports; i++) { 1002ca77329fSKristen Carlson Accardi struct ata_port *ap = host->ports[i]; 1003ca77329fSKristen Carlson Accardi ata_lpm_schedule(ap, ap->pm_policy); 1004ca77329fSKristen Carlson Accardi } 1005ca77329fSKristen Carlson Accardi } 10061992a5edSStephen Rothwell #endif /* CONFIG_PM */ 1007ca77329fSKristen Carlson Accardi 1008ca77329fSKristen Carlson Accardi 1009c6fd2807SJeff Garzik /** 1010c6fd2807SJeff Garzik * ata_devchk - PATA device presence detection 1011c6fd2807SJeff Garzik * @ap: ATA channel to examine 1012c6fd2807SJeff Garzik * @device: Device to examine (starting at zero) 1013c6fd2807SJeff Garzik * 10140d5ff566STejun Heo * This technique was originally described in 10150d5ff566STejun Heo * Hale Landis's ATADRVR (www.ata-atapi.com), and 10160d5ff566STejun Heo * later found its way into the ATA/ATAPI spec. 10170d5ff566STejun Heo * 10180d5ff566STejun Heo * Write a pattern to the ATA shadow registers, 10190d5ff566STejun Heo * and if a device is present, it will respond by 10200d5ff566STejun Heo * correctly storing and echoing back the 10210d5ff566STejun Heo * ATA shadow register contents. 1022c6fd2807SJeff Garzik * 1023c6fd2807SJeff Garzik * LOCKING: 1024c6fd2807SJeff Garzik * caller. 1025c6fd2807SJeff Garzik */ 1026c6fd2807SJeff Garzik 10270d5ff566STejun Heo static unsigned int ata_devchk(struct ata_port *ap, unsigned int device) 1028c6fd2807SJeff Garzik { 10290d5ff566STejun Heo struct ata_ioports *ioaddr = &ap->ioaddr; 10300d5ff566STejun Heo u8 nsect, lbal; 10310d5ff566STejun Heo 10320d5ff566STejun Heo ap->ops->dev_select(ap, device); 10330d5ff566STejun Heo 10340d5ff566STejun Heo iowrite8(0x55, ioaddr->nsect_addr); 10350d5ff566STejun Heo iowrite8(0xaa, ioaddr->lbal_addr); 10360d5ff566STejun Heo 10370d5ff566STejun Heo iowrite8(0xaa, ioaddr->nsect_addr); 10380d5ff566STejun Heo iowrite8(0x55, ioaddr->lbal_addr); 10390d5ff566STejun Heo 10400d5ff566STejun Heo iowrite8(0x55, ioaddr->nsect_addr); 10410d5ff566STejun Heo iowrite8(0xaa, ioaddr->lbal_addr); 10420d5ff566STejun Heo 10430d5ff566STejun Heo nsect = ioread8(ioaddr->nsect_addr); 10440d5ff566STejun Heo lbal = ioread8(ioaddr->lbal_addr); 10450d5ff566STejun Heo 10460d5ff566STejun Heo if ((nsect == 0x55) && (lbal == 0xaa)) 10470d5ff566STejun Heo return 1; /* we found a device */ 10480d5ff566STejun Heo 10490d5ff566STejun Heo return 0; /* nothing found */ 1050c6fd2807SJeff Garzik } 1051c6fd2807SJeff Garzik 1052c6fd2807SJeff Garzik /** 1053c6fd2807SJeff Garzik * ata_dev_classify - determine device type based on ATA-spec signature 1054c6fd2807SJeff Garzik * @tf: ATA taskfile register set for device to be identified 1055c6fd2807SJeff Garzik * 1056c6fd2807SJeff Garzik * Determine from taskfile register contents whether a device is 1057c6fd2807SJeff Garzik * ATA or ATAPI, as per "Signature and persistence" section 1058c6fd2807SJeff Garzik * of ATA/PI spec (volume 1, sect 5.14). 1059c6fd2807SJeff Garzik * 1060c6fd2807SJeff Garzik * LOCKING: 1061c6fd2807SJeff Garzik * None. 1062c6fd2807SJeff Garzik * 1063c6fd2807SJeff Garzik * RETURNS: 1064633273a3STejun Heo * Device type, %ATA_DEV_ATA, %ATA_DEV_ATAPI, %ATA_DEV_PMP or 1065633273a3STejun Heo * %ATA_DEV_UNKNOWN the event of failure. 1066c6fd2807SJeff Garzik */ 1067c6fd2807SJeff Garzik unsigned int ata_dev_classify(const struct ata_taskfile *tf) 1068c6fd2807SJeff Garzik { 1069c6fd2807SJeff Garzik /* Apple's open source Darwin code hints that some devices only 1070c6fd2807SJeff Garzik * put a proper signature into the LBA mid/high registers, 1071c6fd2807SJeff Garzik * So, we only check those. It's sufficient for uniqueness. 1072633273a3STejun Heo * 1073633273a3STejun Heo * ATA/ATAPI-7 (d1532v1r1: Feb. 19, 2003) specified separate 1074633273a3STejun Heo * signatures for ATA and ATAPI devices attached on SerialATA, 1075633273a3STejun Heo * 0x3c/0xc3 and 0x69/0x96 respectively. However, SerialATA 1076633273a3STejun Heo * spec has never mentioned about using different signatures 1077633273a3STejun Heo * for ATA/ATAPI devices. Then, Serial ATA II: Port 1078633273a3STejun Heo * Multiplier specification began to use 0x69/0x96 to identify 1079633273a3STejun Heo * port multpliers and 0x3c/0xc3 to identify SEMB device. 1080633273a3STejun Heo * ATA/ATAPI-7 dropped descriptions about 0x3c/0xc3 and 1081633273a3STejun Heo * 0x69/0x96 shortly and described them as reserved for 1082633273a3STejun Heo * SerialATA. 1083633273a3STejun Heo * 1084633273a3STejun Heo * We follow the current spec and consider that 0x69/0x96 1085633273a3STejun Heo * identifies a port multiplier and 0x3c/0xc3 a SEMB device. 1086c6fd2807SJeff Garzik */ 1087633273a3STejun Heo if ((tf->lbam == 0) && (tf->lbah == 0)) { 1088c6fd2807SJeff Garzik DPRINTK("found ATA device by sig\n"); 1089c6fd2807SJeff Garzik return ATA_DEV_ATA; 1090c6fd2807SJeff Garzik } 1091c6fd2807SJeff Garzik 1092633273a3STejun Heo if ((tf->lbam == 0x14) && (tf->lbah == 0xeb)) { 1093c6fd2807SJeff Garzik DPRINTK("found ATAPI device by sig\n"); 1094c6fd2807SJeff Garzik return ATA_DEV_ATAPI; 1095c6fd2807SJeff Garzik } 1096c6fd2807SJeff Garzik 1097633273a3STejun Heo if ((tf->lbam == 0x69) && (tf->lbah == 0x96)) { 1098633273a3STejun Heo DPRINTK("found PMP device by sig\n"); 1099633273a3STejun Heo return ATA_DEV_PMP; 1100633273a3STejun Heo } 1101633273a3STejun Heo 1102633273a3STejun Heo if ((tf->lbam == 0x3c) && (tf->lbah == 0xc3)) { 11032dcb407eSJeff Garzik printk(KERN_INFO "ata: SEMB device ignored\n"); 1104633273a3STejun Heo return ATA_DEV_SEMB_UNSUP; /* not yet */ 1105633273a3STejun Heo } 1106633273a3STejun Heo 1107c6fd2807SJeff Garzik DPRINTK("unknown device\n"); 1108c6fd2807SJeff Garzik return ATA_DEV_UNKNOWN; 1109c6fd2807SJeff Garzik } 1110c6fd2807SJeff Garzik 1111c6fd2807SJeff Garzik /** 1112c6fd2807SJeff Garzik * ata_dev_try_classify - Parse returned ATA device signature 11133f19859eSTejun Heo * @dev: ATA device to classify (starting at zero) 11143f19859eSTejun Heo * @present: device seems present 1115c6fd2807SJeff Garzik * @r_err: Value of error register on completion 1116c6fd2807SJeff Garzik * 1117c6fd2807SJeff Garzik * After an event -- SRST, E.D.D., or SATA COMRESET -- occurs, 1118c6fd2807SJeff Garzik * an ATA/ATAPI-defined set of values is placed in the ATA 1119c6fd2807SJeff Garzik * shadow registers, indicating the results of device detection 1120c6fd2807SJeff Garzik * and diagnostics. 1121c6fd2807SJeff Garzik * 1122c6fd2807SJeff Garzik * Select the ATA device, and read the values from the ATA shadow 1123c6fd2807SJeff Garzik * registers. Then parse according to the Error register value, 1124c6fd2807SJeff Garzik * and the spec-defined values examined by ata_dev_classify(). 1125c6fd2807SJeff Garzik * 1126c6fd2807SJeff Garzik * LOCKING: 1127c6fd2807SJeff Garzik * caller. 1128c6fd2807SJeff Garzik * 1129c6fd2807SJeff Garzik * RETURNS: 1130c6fd2807SJeff Garzik * Device type - %ATA_DEV_ATA, %ATA_DEV_ATAPI or %ATA_DEV_NONE. 1131c6fd2807SJeff Garzik */ 11323f19859eSTejun Heo unsigned int ata_dev_try_classify(struct ata_device *dev, int present, 11333f19859eSTejun Heo u8 *r_err) 1134c6fd2807SJeff Garzik { 11353f19859eSTejun Heo struct ata_port *ap = dev->link->ap; 1136c6fd2807SJeff Garzik struct ata_taskfile tf; 1137c6fd2807SJeff Garzik unsigned int class; 1138c6fd2807SJeff Garzik u8 err; 1139c6fd2807SJeff Garzik 11403f19859eSTejun Heo ap->ops->dev_select(ap, dev->devno); 1141c6fd2807SJeff Garzik 1142c6fd2807SJeff Garzik memset(&tf, 0, sizeof(tf)); 1143c6fd2807SJeff Garzik 1144c6fd2807SJeff Garzik ap->ops->tf_read(ap, &tf); 1145c6fd2807SJeff Garzik err = tf.feature; 1146c6fd2807SJeff Garzik if (r_err) 1147c6fd2807SJeff Garzik *r_err = err; 1148c6fd2807SJeff Garzik 1149c5038fc0SAlan Cox /* see if device passed diags: continue and warn later */ 1150c5038fc0SAlan Cox if (err == 0) 115193590859SAlan Cox /* diagnostic fail : do nothing _YET_ */ 11523f19859eSTejun Heo dev->horkage |= ATA_HORKAGE_DIAGNOSTIC; 115393590859SAlan Cox else if (err == 1) 1154c6fd2807SJeff Garzik /* do nothing */ ; 11553f19859eSTejun Heo else if ((dev->devno == 0) && (err == 0x81)) 1156c6fd2807SJeff Garzik /* do nothing */ ; 1157c6fd2807SJeff Garzik else 1158c6fd2807SJeff Garzik return ATA_DEV_NONE; 1159c6fd2807SJeff Garzik 1160c6fd2807SJeff Garzik /* determine if device is ATA or ATAPI */ 1161c6fd2807SJeff Garzik class = ata_dev_classify(&tf); 1162c6fd2807SJeff Garzik 1163d7fbee05STejun Heo if (class == ATA_DEV_UNKNOWN) { 1164d7fbee05STejun Heo /* If the device failed diagnostic, it's likely to 1165d7fbee05STejun Heo * have reported incorrect device signature too. 1166d7fbee05STejun Heo * Assume ATA device if the device seems present but 1167d7fbee05STejun Heo * device signature is invalid with diagnostic 1168d7fbee05STejun Heo * failure. 1169d7fbee05STejun Heo */ 1170d7fbee05STejun Heo if (present && (dev->horkage & ATA_HORKAGE_DIAGNOSTIC)) 1171d7fbee05STejun Heo class = ATA_DEV_ATA; 1172d7fbee05STejun Heo else 1173d7fbee05STejun Heo class = ATA_DEV_NONE; 1174d7fbee05STejun Heo } else if ((class == ATA_DEV_ATA) && (ata_chk_status(ap) == 0)) 1175d7fbee05STejun Heo class = ATA_DEV_NONE; 1176d7fbee05STejun Heo 1177c6fd2807SJeff Garzik return class; 1178c6fd2807SJeff Garzik } 1179c6fd2807SJeff Garzik 1180c6fd2807SJeff Garzik /** 1181c6fd2807SJeff Garzik * ata_id_string - Convert IDENTIFY DEVICE page into string 1182c6fd2807SJeff Garzik * @id: IDENTIFY DEVICE results we will examine 1183c6fd2807SJeff Garzik * @s: string into which data is output 1184c6fd2807SJeff Garzik * @ofs: offset into identify device page 1185c6fd2807SJeff Garzik * @len: length of string to return. must be an even number. 1186c6fd2807SJeff Garzik * 1187c6fd2807SJeff Garzik * The strings in the IDENTIFY DEVICE page are broken up into 1188c6fd2807SJeff Garzik * 16-bit chunks. Run through the string, and output each 1189c6fd2807SJeff Garzik * 8-bit chunk linearly, regardless of platform. 1190c6fd2807SJeff Garzik * 1191c6fd2807SJeff Garzik * LOCKING: 1192c6fd2807SJeff Garzik * caller. 1193c6fd2807SJeff Garzik */ 1194c6fd2807SJeff Garzik 1195c6fd2807SJeff Garzik void ata_id_string(const u16 *id, unsigned char *s, 1196c6fd2807SJeff Garzik unsigned int ofs, unsigned int len) 1197c6fd2807SJeff Garzik { 1198c6fd2807SJeff Garzik unsigned int c; 1199c6fd2807SJeff Garzik 1200c6fd2807SJeff Garzik while (len > 0) { 1201c6fd2807SJeff Garzik c = id[ofs] >> 8; 1202c6fd2807SJeff Garzik *s = c; 1203c6fd2807SJeff Garzik s++; 1204c6fd2807SJeff Garzik 1205c6fd2807SJeff Garzik c = id[ofs] & 0xff; 1206c6fd2807SJeff Garzik *s = c; 1207c6fd2807SJeff Garzik s++; 1208c6fd2807SJeff Garzik 1209c6fd2807SJeff Garzik ofs++; 1210c6fd2807SJeff Garzik len -= 2; 1211c6fd2807SJeff Garzik } 1212c6fd2807SJeff Garzik } 1213c6fd2807SJeff Garzik 1214c6fd2807SJeff Garzik /** 1215c6fd2807SJeff Garzik * ata_id_c_string - Convert IDENTIFY DEVICE page into C string 1216c6fd2807SJeff Garzik * @id: IDENTIFY DEVICE results we will examine 1217c6fd2807SJeff Garzik * @s: string into which data is output 1218c6fd2807SJeff Garzik * @ofs: offset into identify device page 1219c6fd2807SJeff Garzik * @len: length of string to return. must be an odd number. 1220c6fd2807SJeff Garzik * 1221c6fd2807SJeff Garzik * This function is identical to ata_id_string except that it 1222c6fd2807SJeff Garzik * trims trailing spaces and terminates the resulting string with 1223c6fd2807SJeff Garzik * null. @len must be actual maximum length (even number) + 1. 1224c6fd2807SJeff Garzik * 1225c6fd2807SJeff Garzik * LOCKING: 1226c6fd2807SJeff Garzik * caller. 1227c6fd2807SJeff Garzik */ 1228c6fd2807SJeff Garzik void ata_id_c_string(const u16 *id, unsigned char *s, 1229c6fd2807SJeff Garzik unsigned int ofs, unsigned int len) 1230c6fd2807SJeff Garzik { 1231c6fd2807SJeff Garzik unsigned char *p; 1232c6fd2807SJeff Garzik 1233c6fd2807SJeff Garzik WARN_ON(!(len & 1)); 1234c6fd2807SJeff Garzik 1235c6fd2807SJeff Garzik ata_id_string(id, s, ofs, len - 1); 1236c6fd2807SJeff Garzik 1237c6fd2807SJeff Garzik p = s + strnlen(s, len - 1); 1238c6fd2807SJeff Garzik while (p > s && p[-1] == ' ') 1239c6fd2807SJeff Garzik p--; 1240c6fd2807SJeff Garzik *p = '\0'; 1241c6fd2807SJeff Garzik } 1242c6fd2807SJeff Garzik 1243db6f8759STejun Heo static u64 ata_id_n_sectors(const u16 *id) 1244db6f8759STejun Heo { 1245db6f8759STejun Heo if (ata_id_has_lba(id)) { 1246db6f8759STejun Heo if (ata_id_has_lba48(id)) 1247db6f8759STejun Heo return ata_id_u64(id, 100); 1248db6f8759STejun Heo else 1249db6f8759STejun Heo return ata_id_u32(id, 60); 1250db6f8759STejun Heo } else { 1251db6f8759STejun Heo if (ata_id_current_chs_valid(id)) 1252db6f8759STejun Heo return ata_id_u32(id, 57); 1253db6f8759STejun Heo else 1254db6f8759STejun Heo return id[1] * id[3] * id[6]; 1255db6f8759STejun Heo } 1256db6f8759STejun Heo } 1257db6f8759STejun Heo 12581e999736SAlan Cox static u64 ata_tf_to_lba48(struct ata_taskfile *tf) 12591e999736SAlan Cox { 12601e999736SAlan Cox u64 sectors = 0; 12611e999736SAlan Cox 12621e999736SAlan Cox sectors |= ((u64)(tf->hob_lbah & 0xff)) << 40; 12631e999736SAlan Cox sectors |= ((u64)(tf->hob_lbam & 0xff)) << 32; 12641e999736SAlan Cox sectors |= (tf->hob_lbal & 0xff) << 24; 12651e999736SAlan Cox sectors |= (tf->lbah & 0xff) << 16; 12661e999736SAlan Cox sectors |= (tf->lbam & 0xff) << 8; 12671e999736SAlan Cox sectors |= (tf->lbal & 0xff); 12681e999736SAlan Cox 12691e999736SAlan Cox return ++sectors; 12701e999736SAlan Cox } 12711e999736SAlan Cox 12721e999736SAlan Cox static u64 ata_tf_to_lba(struct ata_taskfile *tf) 12731e999736SAlan Cox { 12741e999736SAlan Cox u64 sectors = 0; 12751e999736SAlan Cox 12761e999736SAlan Cox sectors |= (tf->device & 0x0f) << 24; 12771e999736SAlan Cox sectors |= (tf->lbah & 0xff) << 16; 12781e999736SAlan Cox sectors |= (tf->lbam & 0xff) << 8; 12791e999736SAlan Cox sectors |= (tf->lbal & 0xff); 12801e999736SAlan Cox 12811e999736SAlan Cox return ++sectors; 12821e999736SAlan Cox } 12831e999736SAlan Cox 12841e999736SAlan Cox /** 1285c728a914STejun Heo * ata_read_native_max_address - Read native max address 1286c728a914STejun Heo * @dev: target device 1287c728a914STejun Heo * @max_sectors: out parameter for the result native max address 12881e999736SAlan Cox * 1289c728a914STejun Heo * Perform an LBA48 or LBA28 native size query upon the device in 1290c728a914STejun Heo * question. 1291c728a914STejun Heo * 1292c728a914STejun Heo * RETURNS: 1293c728a914STejun Heo * 0 on success, -EACCES if command is aborted by the drive. 1294c728a914STejun Heo * -EIO on other errors. 12951e999736SAlan Cox */ 1296c728a914STejun Heo static int ata_read_native_max_address(struct ata_device *dev, u64 *max_sectors) 12971e999736SAlan Cox { 1298c728a914STejun Heo unsigned int err_mask; 12991e999736SAlan Cox struct ata_taskfile tf; 1300c728a914STejun Heo int lba48 = ata_id_has_lba48(dev->id); 13011e999736SAlan Cox 13021e999736SAlan Cox ata_tf_init(dev, &tf); 13031e999736SAlan Cox 1304c728a914STejun Heo /* always clear all address registers */ 13051e999736SAlan Cox tf.flags |= ATA_TFLAG_DEVICE | ATA_TFLAG_ISADDR; 1306c728a914STejun Heo 1307c728a914STejun Heo if (lba48) { 1308c728a914STejun Heo tf.command = ATA_CMD_READ_NATIVE_MAX_EXT; 1309c728a914STejun Heo tf.flags |= ATA_TFLAG_LBA48; 1310c728a914STejun Heo } else 1311c728a914STejun Heo tf.command = ATA_CMD_READ_NATIVE_MAX; 1312c728a914STejun Heo 13131e999736SAlan Cox tf.protocol |= ATA_PROT_NODATA; 1314c728a914STejun Heo tf.device |= ATA_LBA; 13151e999736SAlan Cox 13162b789108STejun Heo err_mask = ata_exec_internal(dev, &tf, NULL, DMA_NONE, NULL, 0, 0); 1317c728a914STejun Heo if (err_mask) { 1318c728a914STejun Heo ata_dev_printk(dev, KERN_WARNING, "failed to read native " 1319c728a914STejun Heo "max address (err_mask=0x%x)\n", err_mask); 1320c728a914STejun Heo if (err_mask == AC_ERR_DEV && (tf.feature & ATA_ABORTED)) 1321c728a914STejun Heo return -EACCES; 1322c728a914STejun Heo return -EIO; 1323c728a914STejun Heo } 1324c728a914STejun Heo 1325c728a914STejun Heo if (lba48) 1326c728a914STejun Heo *max_sectors = ata_tf_to_lba48(&tf); 1327c728a914STejun Heo else 1328c728a914STejun Heo *max_sectors = ata_tf_to_lba(&tf); 132993328e11SAlan Cox if (dev->horkage & ATA_HORKAGE_HPA_SIZE) 133093328e11SAlan Cox (*max_sectors)--; 13311e999736SAlan Cox return 0; 13321e999736SAlan Cox } 13331e999736SAlan Cox 13341e999736SAlan Cox /** 1335c728a914STejun Heo * ata_set_max_sectors - Set max sectors 1336c728a914STejun Heo * @dev: target device 13376b38d1d1SRandy Dunlap * @new_sectors: new max sectors value to set for the device 13381e999736SAlan Cox * 1339c728a914STejun Heo * Set max sectors of @dev to @new_sectors. 1340c728a914STejun Heo * 1341c728a914STejun Heo * RETURNS: 1342c728a914STejun Heo * 0 on success, -EACCES if command is aborted or denied (due to 1343c728a914STejun Heo * previous non-volatile SET_MAX) by the drive. -EIO on other 1344c728a914STejun Heo * errors. 13451e999736SAlan Cox */ 134605027adcSTejun Heo static int ata_set_max_sectors(struct ata_device *dev, u64 new_sectors) 13471e999736SAlan Cox { 1348c728a914STejun Heo unsigned int err_mask; 13491e999736SAlan Cox struct ata_taskfile tf; 1350c728a914STejun Heo int lba48 = ata_id_has_lba48(dev->id); 13511e999736SAlan Cox 13521e999736SAlan Cox new_sectors--; 13531e999736SAlan Cox 13541e999736SAlan Cox ata_tf_init(dev, &tf); 13551e999736SAlan Cox 1356c728a914STejun Heo tf.flags |= ATA_TFLAG_DEVICE | ATA_TFLAG_ISADDR; 13571e999736SAlan Cox 1358c728a914STejun Heo if (lba48) { 1359c728a914STejun Heo tf.command = ATA_CMD_SET_MAX_EXT; 1360c728a914STejun Heo tf.flags |= ATA_TFLAG_LBA48; 13611e999736SAlan Cox 13621e999736SAlan Cox tf.hob_lbal = (new_sectors >> 24) & 0xff; 13631e999736SAlan Cox tf.hob_lbam = (new_sectors >> 32) & 0xff; 13641e999736SAlan Cox tf.hob_lbah = (new_sectors >> 40) & 0xff; 13651e582ba4STejun Heo } else { 13661e999736SAlan Cox tf.command = ATA_CMD_SET_MAX; 1367c728a914STejun Heo 13681e582ba4STejun Heo tf.device |= (new_sectors >> 24) & 0xf; 13691e582ba4STejun Heo } 13701e582ba4STejun Heo 13711e999736SAlan Cox tf.protocol |= ATA_PROT_NODATA; 1372c728a914STejun Heo tf.device |= ATA_LBA; 13731e999736SAlan Cox 13741e999736SAlan Cox tf.lbal = (new_sectors >> 0) & 0xff; 13751e999736SAlan Cox tf.lbam = (new_sectors >> 8) & 0xff; 13761e999736SAlan Cox tf.lbah = (new_sectors >> 16) & 0xff; 13771e999736SAlan Cox 13782b789108STejun Heo err_mask = ata_exec_internal(dev, &tf, NULL, DMA_NONE, NULL, 0, 0); 1379c728a914STejun Heo if (err_mask) { 1380c728a914STejun Heo ata_dev_printk(dev, KERN_WARNING, "failed to set " 1381c728a914STejun Heo "max address (err_mask=0x%x)\n", err_mask); 1382c728a914STejun Heo if (err_mask == AC_ERR_DEV && 1383c728a914STejun Heo (tf.feature & (ATA_ABORTED | ATA_IDNF))) 1384c728a914STejun Heo return -EACCES; 1385c728a914STejun Heo return -EIO; 1386c728a914STejun Heo } 1387c728a914STejun Heo 13881e999736SAlan Cox return 0; 13891e999736SAlan Cox } 13901e999736SAlan Cox 13911e999736SAlan Cox /** 13921e999736SAlan Cox * ata_hpa_resize - Resize a device with an HPA set 13931e999736SAlan Cox * @dev: Device to resize 13941e999736SAlan Cox * 13951e999736SAlan Cox * Read the size of an LBA28 or LBA48 disk with HPA features and resize 13961e999736SAlan Cox * it if required to the full size of the media. The caller must check 13971e999736SAlan Cox * the drive has the HPA feature set enabled. 139805027adcSTejun Heo * 139905027adcSTejun Heo * RETURNS: 140005027adcSTejun Heo * 0 on success, -errno on failure. 14011e999736SAlan Cox */ 140205027adcSTejun Heo static int ata_hpa_resize(struct ata_device *dev) 14031e999736SAlan Cox { 140405027adcSTejun Heo struct ata_eh_context *ehc = &dev->link->eh_context; 140505027adcSTejun Heo int print_info = ehc->i.flags & ATA_EHI_PRINTINFO; 140605027adcSTejun Heo u64 sectors = ata_id_n_sectors(dev->id); 140705027adcSTejun Heo u64 native_sectors; 1408c728a914STejun Heo int rc; 14091e999736SAlan Cox 141005027adcSTejun Heo /* do we need to do it? */ 141105027adcSTejun Heo if (dev->class != ATA_DEV_ATA || 141205027adcSTejun Heo !ata_id_has_lba(dev->id) || !ata_id_hpa_enabled(dev->id) || 141305027adcSTejun Heo (dev->horkage & ATA_HORKAGE_BROKEN_HPA)) 1414c728a914STejun Heo return 0; 14151e999736SAlan Cox 141605027adcSTejun Heo /* read native max address */ 141705027adcSTejun Heo rc = ata_read_native_max_address(dev, &native_sectors); 141805027adcSTejun Heo if (rc) { 141905027adcSTejun Heo /* If HPA isn't going to be unlocked, skip HPA 142005027adcSTejun Heo * resizing from the next try. 142105027adcSTejun Heo */ 142205027adcSTejun Heo if (!ata_ignore_hpa) { 142305027adcSTejun Heo ata_dev_printk(dev, KERN_WARNING, "HPA support seems " 142405027adcSTejun Heo "broken, will skip HPA handling\n"); 142505027adcSTejun Heo dev->horkage |= ATA_HORKAGE_BROKEN_HPA; 142605027adcSTejun Heo 142705027adcSTejun Heo /* we can continue if device aborted the command */ 142805027adcSTejun Heo if (rc == -EACCES) 142905027adcSTejun Heo rc = 0; 143005027adcSTejun Heo } 143105027adcSTejun Heo 143205027adcSTejun Heo return rc; 143305027adcSTejun Heo } 143405027adcSTejun Heo 143505027adcSTejun Heo /* nothing to do? */ 143605027adcSTejun Heo if (native_sectors <= sectors || !ata_ignore_hpa) { 143705027adcSTejun Heo if (!print_info || native_sectors == sectors) 143805027adcSTejun Heo return 0; 143905027adcSTejun Heo 144005027adcSTejun Heo if (native_sectors > sectors) 14411e999736SAlan Cox ata_dev_printk(dev, KERN_INFO, 144205027adcSTejun Heo "HPA detected: current %llu, native %llu\n", 144305027adcSTejun Heo (unsigned long long)sectors, 144405027adcSTejun Heo (unsigned long long)native_sectors); 144505027adcSTejun Heo else if (native_sectors < sectors) 144605027adcSTejun Heo ata_dev_printk(dev, KERN_WARNING, 144705027adcSTejun Heo "native sectors (%llu) is smaller than " 144805027adcSTejun Heo "sectors (%llu)\n", 144905027adcSTejun Heo (unsigned long long)native_sectors, 145005027adcSTejun Heo (unsigned long long)sectors); 145105027adcSTejun Heo return 0; 14521e999736SAlan Cox } 145337301a55STejun Heo 145405027adcSTejun Heo /* let's unlock HPA */ 145505027adcSTejun Heo rc = ata_set_max_sectors(dev, native_sectors); 145605027adcSTejun Heo if (rc == -EACCES) { 145705027adcSTejun Heo /* if device aborted the command, skip HPA resizing */ 145805027adcSTejun Heo ata_dev_printk(dev, KERN_WARNING, "device aborted resize " 145905027adcSTejun Heo "(%llu -> %llu), skipping HPA handling\n", 146005027adcSTejun Heo (unsigned long long)sectors, 146105027adcSTejun Heo (unsigned long long)native_sectors); 146205027adcSTejun Heo dev->horkage |= ATA_HORKAGE_BROKEN_HPA; 146305027adcSTejun Heo return 0; 146405027adcSTejun Heo } else if (rc) 146505027adcSTejun Heo return rc; 146605027adcSTejun Heo 146705027adcSTejun Heo /* re-read IDENTIFY data */ 146805027adcSTejun Heo rc = ata_dev_reread_id(dev, 0); 146905027adcSTejun Heo if (rc) { 147005027adcSTejun Heo ata_dev_printk(dev, KERN_ERR, "failed to re-read IDENTIFY " 147105027adcSTejun Heo "data after HPA resizing\n"); 147205027adcSTejun Heo return rc; 147305027adcSTejun Heo } 147405027adcSTejun Heo 147505027adcSTejun Heo if (print_info) { 147605027adcSTejun Heo u64 new_sectors = ata_id_n_sectors(dev->id); 147705027adcSTejun Heo ata_dev_printk(dev, KERN_INFO, 147805027adcSTejun Heo "HPA unlocked: %llu -> %llu, native %llu\n", 147905027adcSTejun Heo (unsigned long long)sectors, 148005027adcSTejun Heo (unsigned long long)new_sectors, 148105027adcSTejun Heo (unsigned long long)native_sectors); 148205027adcSTejun Heo } 148305027adcSTejun Heo 148405027adcSTejun Heo return 0; 14851e999736SAlan Cox } 14861e999736SAlan Cox 1487c6fd2807SJeff Garzik /** 1488c6fd2807SJeff Garzik * ata_noop_dev_select - Select device 0/1 on ATA bus 1489c6fd2807SJeff Garzik * @ap: ATA channel to manipulate 1490c6fd2807SJeff Garzik * @device: ATA device (numbered from zero) to select 1491c6fd2807SJeff Garzik * 1492c6fd2807SJeff Garzik * This function performs no actual function. 1493c6fd2807SJeff Garzik * 1494c6fd2807SJeff Garzik * May be used as the dev_select() entry in ata_port_operations. 1495c6fd2807SJeff Garzik * 1496c6fd2807SJeff Garzik * LOCKING: 1497c6fd2807SJeff Garzik * caller. 1498c6fd2807SJeff Garzik */ 1499c6fd2807SJeff Garzik void ata_noop_dev_select(struct ata_port *ap, unsigned int device) 1500c6fd2807SJeff Garzik { 1501c6fd2807SJeff Garzik } 1502c6fd2807SJeff Garzik 1503c6fd2807SJeff Garzik 1504c6fd2807SJeff Garzik /** 1505c6fd2807SJeff Garzik * ata_std_dev_select - Select device 0/1 on ATA bus 1506c6fd2807SJeff Garzik * @ap: ATA channel to manipulate 1507c6fd2807SJeff Garzik * @device: ATA device (numbered from zero) to select 1508c6fd2807SJeff Garzik * 1509c6fd2807SJeff Garzik * Use the method defined in the ATA specification to 1510c6fd2807SJeff Garzik * make either device 0, or device 1, active on the 1511c6fd2807SJeff Garzik * ATA channel. Works with both PIO and MMIO. 1512c6fd2807SJeff Garzik * 1513c6fd2807SJeff Garzik * May be used as the dev_select() entry in ata_port_operations. 1514c6fd2807SJeff Garzik * 1515c6fd2807SJeff Garzik * LOCKING: 1516c6fd2807SJeff Garzik * caller. 1517c6fd2807SJeff Garzik */ 1518c6fd2807SJeff Garzik 1519c6fd2807SJeff Garzik void ata_std_dev_select(struct ata_port *ap, unsigned int device) 1520c6fd2807SJeff Garzik { 1521c6fd2807SJeff Garzik u8 tmp; 1522c6fd2807SJeff Garzik 1523c6fd2807SJeff Garzik if (device == 0) 1524c6fd2807SJeff Garzik tmp = ATA_DEVICE_OBS; 1525c6fd2807SJeff Garzik else 1526c6fd2807SJeff Garzik tmp = ATA_DEVICE_OBS | ATA_DEV1; 1527c6fd2807SJeff Garzik 15280d5ff566STejun Heo iowrite8(tmp, ap->ioaddr.device_addr); 1529c6fd2807SJeff Garzik ata_pause(ap); /* needed; also flushes, for mmio */ 1530c6fd2807SJeff Garzik } 1531c6fd2807SJeff Garzik 1532c6fd2807SJeff Garzik /** 1533c6fd2807SJeff Garzik * ata_dev_select - Select device 0/1 on ATA bus 1534c6fd2807SJeff Garzik * @ap: ATA channel to manipulate 1535c6fd2807SJeff Garzik * @device: ATA device (numbered from zero) to select 1536c6fd2807SJeff Garzik * @wait: non-zero to wait for Status register BSY bit to clear 1537c6fd2807SJeff Garzik * @can_sleep: non-zero if context allows sleeping 1538c6fd2807SJeff Garzik * 1539c6fd2807SJeff Garzik * Use the method defined in the ATA specification to 1540c6fd2807SJeff Garzik * make either device 0, or device 1, active on the 1541c6fd2807SJeff Garzik * ATA channel. 1542c6fd2807SJeff Garzik * 1543c6fd2807SJeff Garzik * This is a high-level version of ata_std_dev_select(), 1544c6fd2807SJeff Garzik * which additionally provides the services of inserting 1545c6fd2807SJeff Garzik * the proper pauses and status polling, where needed. 1546c6fd2807SJeff Garzik * 1547c6fd2807SJeff Garzik * LOCKING: 1548c6fd2807SJeff Garzik * caller. 1549c6fd2807SJeff Garzik */ 1550c6fd2807SJeff Garzik 1551c6fd2807SJeff Garzik void ata_dev_select(struct ata_port *ap, unsigned int device, 1552c6fd2807SJeff Garzik unsigned int wait, unsigned int can_sleep) 1553c6fd2807SJeff Garzik { 1554c6fd2807SJeff Garzik if (ata_msg_probe(ap)) 155544877b4eSTejun Heo ata_port_printk(ap, KERN_INFO, "ata_dev_select: ENTER, " 155644877b4eSTejun Heo "device %u, wait %u\n", device, wait); 1557c6fd2807SJeff Garzik 1558c6fd2807SJeff Garzik if (wait) 1559c6fd2807SJeff Garzik ata_wait_idle(ap); 1560c6fd2807SJeff Garzik 1561c6fd2807SJeff Garzik ap->ops->dev_select(ap, device); 1562c6fd2807SJeff Garzik 1563c6fd2807SJeff Garzik if (wait) { 15649af5c9c9STejun Heo if (can_sleep && ap->link.device[device].class == ATA_DEV_ATAPI) 1565c6fd2807SJeff Garzik msleep(150); 1566c6fd2807SJeff Garzik ata_wait_idle(ap); 1567c6fd2807SJeff Garzik } 1568c6fd2807SJeff Garzik } 1569c6fd2807SJeff Garzik 1570c6fd2807SJeff Garzik /** 1571c6fd2807SJeff Garzik * ata_dump_id - IDENTIFY DEVICE info debugging output 1572c6fd2807SJeff Garzik * @id: IDENTIFY DEVICE page to dump 1573c6fd2807SJeff Garzik * 1574c6fd2807SJeff Garzik * Dump selected 16-bit words from the given IDENTIFY DEVICE 1575c6fd2807SJeff Garzik * page. 1576c6fd2807SJeff Garzik * 1577c6fd2807SJeff Garzik * LOCKING: 1578c6fd2807SJeff Garzik * caller. 1579c6fd2807SJeff Garzik */ 1580c6fd2807SJeff Garzik 1581c6fd2807SJeff Garzik static inline void ata_dump_id(const u16 *id) 1582c6fd2807SJeff Garzik { 1583c6fd2807SJeff Garzik DPRINTK("49==0x%04x " 1584c6fd2807SJeff Garzik "53==0x%04x " 1585c6fd2807SJeff Garzik "63==0x%04x " 1586c6fd2807SJeff Garzik "64==0x%04x " 1587c6fd2807SJeff Garzik "75==0x%04x \n", 1588c6fd2807SJeff Garzik id[49], 1589c6fd2807SJeff Garzik id[53], 1590c6fd2807SJeff Garzik id[63], 1591c6fd2807SJeff Garzik id[64], 1592c6fd2807SJeff Garzik id[75]); 1593c6fd2807SJeff Garzik DPRINTK("80==0x%04x " 1594c6fd2807SJeff Garzik "81==0x%04x " 1595c6fd2807SJeff Garzik "82==0x%04x " 1596c6fd2807SJeff Garzik "83==0x%04x " 1597c6fd2807SJeff Garzik "84==0x%04x \n", 1598c6fd2807SJeff Garzik id[80], 1599c6fd2807SJeff Garzik id[81], 1600c6fd2807SJeff Garzik id[82], 1601c6fd2807SJeff Garzik id[83], 1602c6fd2807SJeff Garzik id[84]); 1603c6fd2807SJeff Garzik DPRINTK("88==0x%04x " 1604c6fd2807SJeff Garzik "93==0x%04x\n", 1605c6fd2807SJeff Garzik id[88], 1606c6fd2807SJeff Garzik id[93]); 1607c6fd2807SJeff Garzik } 1608c6fd2807SJeff Garzik 1609c6fd2807SJeff Garzik /** 1610c6fd2807SJeff Garzik * ata_id_xfermask - Compute xfermask from the given IDENTIFY data 1611c6fd2807SJeff Garzik * @id: IDENTIFY data to compute xfer mask from 1612c6fd2807SJeff Garzik * 1613c6fd2807SJeff Garzik * Compute the xfermask for this device. This is not as trivial 1614c6fd2807SJeff Garzik * as it seems if we must consider early devices correctly. 1615c6fd2807SJeff Garzik * 1616c6fd2807SJeff Garzik * FIXME: pre IDE drive timing (do we care ?). 1617c6fd2807SJeff Garzik * 1618c6fd2807SJeff Garzik * LOCKING: 1619c6fd2807SJeff Garzik * None. 1620c6fd2807SJeff Garzik * 1621c6fd2807SJeff Garzik * RETURNS: 1622c6fd2807SJeff Garzik * Computed xfermask 1623c6fd2807SJeff Garzik */ 16247dc951aeSTejun Heo unsigned long ata_id_xfermask(const u16 *id) 1625c6fd2807SJeff Garzik { 16267dc951aeSTejun Heo unsigned long pio_mask, mwdma_mask, udma_mask; 1627c6fd2807SJeff Garzik 1628c6fd2807SJeff Garzik /* Usual case. Word 53 indicates word 64 is valid */ 1629c6fd2807SJeff Garzik if (id[ATA_ID_FIELD_VALID] & (1 << 1)) { 1630c6fd2807SJeff Garzik pio_mask = id[ATA_ID_PIO_MODES] & 0x03; 1631c6fd2807SJeff Garzik pio_mask <<= 3; 1632c6fd2807SJeff Garzik pio_mask |= 0x7; 1633c6fd2807SJeff Garzik } else { 1634c6fd2807SJeff Garzik /* If word 64 isn't valid then Word 51 high byte holds 1635c6fd2807SJeff Garzik * the PIO timing number for the maximum. Turn it into 1636c6fd2807SJeff Garzik * a mask. 1637c6fd2807SJeff Garzik */ 16387a0f1c8aSLennert Buytenhek u8 mode = (id[ATA_ID_OLD_PIO_MODES] >> 8) & 0xFF; 163946767aebSAlan Cox if (mode < 5) /* Valid PIO range */ 164046767aebSAlan Cox pio_mask = (2 << mode) - 1; 164146767aebSAlan Cox else 164246767aebSAlan Cox pio_mask = 1; 1643c6fd2807SJeff Garzik 1644c6fd2807SJeff Garzik /* But wait.. there's more. Design your standards by 1645c6fd2807SJeff Garzik * committee and you too can get a free iordy field to 1646c6fd2807SJeff Garzik * process. However its the speeds not the modes that 1647c6fd2807SJeff Garzik * are supported... Note drivers using the timing API 1648c6fd2807SJeff Garzik * will get this right anyway 1649c6fd2807SJeff Garzik */ 1650c6fd2807SJeff Garzik } 1651c6fd2807SJeff Garzik 1652c6fd2807SJeff Garzik mwdma_mask = id[ATA_ID_MWDMA_MODES] & 0x07; 1653c6fd2807SJeff Garzik 1654b352e57dSAlan Cox if (ata_id_is_cfa(id)) { 1655b352e57dSAlan Cox /* 1656b352e57dSAlan Cox * Process compact flash extended modes 1657b352e57dSAlan Cox */ 1658b352e57dSAlan Cox int pio = id[163] & 0x7; 1659b352e57dSAlan Cox int dma = (id[163] >> 3) & 7; 1660b352e57dSAlan Cox 1661b352e57dSAlan Cox if (pio) 1662b352e57dSAlan Cox pio_mask |= (1 << 5); 1663b352e57dSAlan Cox if (pio > 1) 1664b352e57dSAlan Cox pio_mask |= (1 << 6); 1665b352e57dSAlan Cox if (dma) 1666b352e57dSAlan Cox mwdma_mask |= (1 << 3); 1667b352e57dSAlan Cox if (dma > 1) 1668b352e57dSAlan Cox mwdma_mask |= (1 << 4); 1669b352e57dSAlan Cox } 1670b352e57dSAlan Cox 1671c6fd2807SJeff Garzik udma_mask = 0; 1672c6fd2807SJeff Garzik if (id[ATA_ID_FIELD_VALID] & (1 << 2)) 1673c6fd2807SJeff Garzik udma_mask = id[ATA_ID_UDMA_MODES] & 0xff; 1674c6fd2807SJeff Garzik 1675c6fd2807SJeff Garzik return ata_pack_xfermask(pio_mask, mwdma_mask, udma_mask); 1676c6fd2807SJeff Garzik } 1677c6fd2807SJeff Garzik 1678c6fd2807SJeff Garzik /** 1679442eacc3SJeff Garzik * ata_pio_queue_task - Queue port_task 1680c6fd2807SJeff Garzik * @ap: The ata_port to queue port_task for 1681c6fd2807SJeff Garzik * @fn: workqueue function to be scheduled 168265f27f38SDavid Howells * @data: data for @fn to use 1683c6fd2807SJeff Garzik * @delay: delay time for workqueue function 1684c6fd2807SJeff Garzik * 1685c6fd2807SJeff Garzik * Schedule @fn(@data) for execution after @delay jiffies using 1686c6fd2807SJeff Garzik * port_task. There is one port_task per port and it's the 1687c6fd2807SJeff Garzik * user(low level driver)'s responsibility to make sure that only 1688c6fd2807SJeff Garzik * one task is active at any given time. 1689c6fd2807SJeff Garzik * 1690c6fd2807SJeff Garzik * libata core layer takes care of synchronization between 1691442eacc3SJeff Garzik * port_task and EH. ata_pio_queue_task() may be ignored for EH 1692c6fd2807SJeff Garzik * synchronization. 1693c6fd2807SJeff Garzik * 1694c6fd2807SJeff Garzik * LOCKING: 1695c6fd2807SJeff Garzik * Inherited from caller. 1696c6fd2807SJeff Garzik */ 1697442eacc3SJeff Garzik static void ata_pio_queue_task(struct ata_port *ap, void *data, 1698c6fd2807SJeff Garzik unsigned long delay) 1699c6fd2807SJeff Garzik { 170065f27f38SDavid Howells ap->port_task_data = data; 1701c6fd2807SJeff Garzik 170245a66c1cSOleg Nesterov /* may fail if ata_port_flush_task() in progress */ 170345a66c1cSOleg Nesterov queue_delayed_work(ata_wq, &ap->port_task, delay); 1704c6fd2807SJeff Garzik } 1705c6fd2807SJeff Garzik 1706c6fd2807SJeff Garzik /** 1707c6fd2807SJeff Garzik * ata_port_flush_task - Flush port_task 1708c6fd2807SJeff Garzik * @ap: The ata_port to flush port_task for 1709c6fd2807SJeff Garzik * 1710c6fd2807SJeff Garzik * After this function completes, port_task is guranteed not to 1711c6fd2807SJeff Garzik * be running or scheduled. 1712c6fd2807SJeff Garzik * 1713c6fd2807SJeff Garzik * LOCKING: 1714c6fd2807SJeff Garzik * Kernel thread context (may sleep) 1715c6fd2807SJeff Garzik */ 1716c6fd2807SJeff Garzik void ata_port_flush_task(struct ata_port *ap) 1717c6fd2807SJeff Garzik { 1718c6fd2807SJeff Garzik DPRINTK("ENTER\n"); 1719c6fd2807SJeff Garzik 172045a66c1cSOleg Nesterov cancel_rearming_delayed_work(&ap->port_task); 1721c6fd2807SJeff Garzik 1722c6fd2807SJeff Garzik if (ata_msg_ctl(ap)) 17237f5e4e8dSHarvey Harrison ata_port_printk(ap, KERN_DEBUG, "%s: EXIT\n", __func__); 1724c6fd2807SJeff Garzik } 1725c6fd2807SJeff Garzik 17267102d230SAdrian Bunk static void ata_qc_complete_internal(struct ata_queued_cmd *qc) 1727c6fd2807SJeff Garzik { 1728c6fd2807SJeff Garzik struct completion *waiting = qc->private_data; 1729c6fd2807SJeff Garzik 1730c6fd2807SJeff Garzik complete(waiting); 1731c6fd2807SJeff Garzik } 1732c6fd2807SJeff Garzik 1733c6fd2807SJeff Garzik /** 17342432697bSTejun Heo * ata_exec_internal_sg - execute libata internal command 1735c6fd2807SJeff Garzik * @dev: Device to which the command is sent 1736c6fd2807SJeff Garzik * @tf: Taskfile registers for the command and the result 1737c6fd2807SJeff Garzik * @cdb: CDB for packet command 1738c6fd2807SJeff Garzik * @dma_dir: Data tranfer direction of the command 17395c1ad8b3SRandy Dunlap * @sgl: sg list for the data buffer of the command 17402432697bSTejun Heo * @n_elem: Number of sg entries 17412b789108STejun Heo * @timeout: Timeout in msecs (0 for default) 1742c6fd2807SJeff Garzik * 1743c6fd2807SJeff Garzik * Executes libata internal command with timeout. @tf contains 1744c6fd2807SJeff Garzik * command on entry and result on return. Timeout and error 1745c6fd2807SJeff Garzik * conditions are reported via return value. No recovery action 1746c6fd2807SJeff Garzik * is taken after a command times out. It's caller's duty to 1747c6fd2807SJeff Garzik * clean up after timeout. 1748c6fd2807SJeff Garzik * 1749c6fd2807SJeff Garzik * LOCKING: 1750c6fd2807SJeff Garzik * None. Should be called with kernel context, might sleep. 1751c6fd2807SJeff Garzik * 1752c6fd2807SJeff Garzik * RETURNS: 1753c6fd2807SJeff Garzik * Zero on success, AC_ERR_* mask on failure 1754c6fd2807SJeff Garzik */ 17552432697bSTejun Heo unsigned ata_exec_internal_sg(struct ata_device *dev, 1756c6fd2807SJeff Garzik struct ata_taskfile *tf, const u8 *cdb, 175787260216SJens Axboe int dma_dir, struct scatterlist *sgl, 17582b789108STejun Heo unsigned int n_elem, unsigned long timeout) 1759c6fd2807SJeff Garzik { 17609af5c9c9STejun Heo struct ata_link *link = dev->link; 17619af5c9c9STejun Heo struct ata_port *ap = link->ap; 1762c6fd2807SJeff Garzik u8 command = tf->command; 1763c6fd2807SJeff Garzik struct ata_queued_cmd *qc; 1764c6fd2807SJeff Garzik unsigned int tag, preempted_tag; 1765c6fd2807SJeff Garzik u32 preempted_sactive, preempted_qc_active; 1766da917d69STejun Heo int preempted_nr_active_links; 1767c6fd2807SJeff Garzik DECLARE_COMPLETION_ONSTACK(wait); 1768c6fd2807SJeff Garzik unsigned long flags; 1769c6fd2807SJeff Garzik unsigned int err_mask; 1770c6fd2807SJeff Garzik int rc; 1771c6fd2807SJeff Garzik 1772c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 1773c6fd2807SJeff Garzik 1774c6fd2807SJeff Garzik /* no internal command while frozen */ 1775c6fd2807SJeff Garzik if (ap->pflags & ATA_PFLAG_FROZEN) { 1776c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 1777c6fd2807SJeff Garzik return AC_ERR_SYSTEM; 1778c6fd2807SJeff Garzik } 1779c6fd2807SJeff Garzik 1780c6fd2807SJeff Garzik /* initialize internal qc */ 1781c6fd2807SJeff Garzik 1782c6fd2807SJeff Garzik /* XXX: Tag 0 is used for drivers with legacy EH as some 1783c6fd2807SJeff Garzik * drivers choke if any other tag is given. This breaks 1784c6fd2807SJeff Garzik * ata_tag_internal() test for those drivers. Don't use new 1785c6fd2807SJeff Garzik * EH stuff without converting to it. 1786c6fd2807SJeff Garzik */ 1787c6fd2807SJeff Garzik if (ap->ops->error_handler) 1788c6fd2807SJeff Garzik tag = ATA_TAG_INTERNAL; 1789c6fd2807SJeff Garzik else 1790c6fd2807SJeff Garzik tag = 0; 1791c6fd2807SJeff Garzik 1792c6fd2807SJeff Garzik if (test_and_set_bit(tag, &ap->qc_allocated)) 1793c6fd2807SJeff Garzik BUG(); 1794c6fd2807SJeff Garzik qc = __ata_qc_from_tag(ap, tag); 1795c6fd2807SJeff Garzik 1796c6fd2807SJeff Garzik qc->tag = tag; 1797c6fd2807SJeff Garzik qc->scsicmd = NULL; 1798c6fd2807SJeff Garzik qc->ap = ap; 1799c6fd2807SJeff Garzik qc->dev = dev; 1800c6fd2807SJeff Garzik ata_qc_reinit(qc); 1801c6fd2807SJeff Garzik 18029af5c9c9STejun Heo preempted_tag = link->active_tag; 18039af5c9c9STejun Heo preempted_sactive = link->sactive; 1804c6fd2807SJeff Garzik preempted_qc_active = ap->qc_active; 1805da917d69STejun Heo preempted_nr_active_links = ap->nr_active_links; 18069af5c9c9STejun Heo link->active_tag = ATA_TAG_POISON; 18079af5c9c9STejun Heo link->sactive = 0; 1808c6fd2807SJeff Garzik ap->qc_active = 0; 1809da917d69STejun Heo ap->nr_active_links = 0; 1810c6fd2807SJeff Garzik 1811c6fd2807SJeff Garzik /* prepare & issue qc */ 1812c6fd2807SJeff Garzik qc->tf = *tf; 1813c6fd2807SJeff Garzik if (cdb) 1814c6fd2807SJeff Garzik memcpy(qc->cdb, cdb, ATAPI_CDB_LEN); 1815c6fd2807SJeff Garzik qc->flags |= ATA_QCFLAG_RESULT_TF; 1816c6fd2807SJeff Garzik qc->dma_dir = dma_dir; 1817c6fd2807SJeff Garzik if (dma_dir != DMA_NONE) { 18182432697bSTejun Heo unsigned int i, buflen = 0; 181987260216SJens Axboe struct scatterlist *sg; 18202432697bSTejun Heo 182187260216SJens Axboe for_each_sg(sgl, sg, n_elem, i) 182287260216SJens Axboe buflen += sg->length; 18232432697bSTejun Heo 182487260216SJens Axboe ata_sg_init(qc, sgl, n_elem); 182549c80429SBrian King qc->nbytes = buflen; 1826c6fd2807SJeff Garzik } 1827c6fd2807SJeff Garzik 1828c6fd2807SJeff Garzik qc->private_data = &wait; 1829c6fd2807SJeff Garzik qc->complete_fn = ata_qc_complete_internal; 1830c6fd2807SJeff Garzik 1831c6fd2807SJeff Garzik ata_qc_issue(qc); 1832c6fd2807SJeff Garzik 1833c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 1834c6fd2807SJeff Garzik 18352b789108STejun Heo if (!timeout) 18362b789108STejun Heo timeout = ata_probe_timeout * 1000 / HZ; 18372b789108STejun Heo 18382b789108STejun Heo rc = wait_for_completion_timeout(&wait, msecs_to_jiffies(timeout)); 1839c6fd2807SJeff Garzik 1840c6fd2807SJeff Garzik ata_port_flush_task(ap); 1841c6fd2807SJeff Garzik 1842c6fd2807SJeff Garzik if (!rc) { 1843c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 1844c6fd2807SJeff Garzik 1845c6fd2807SJeff Garzik /* We're racing with irq here. If we lose, the 1846c6fd2807SJeff Garzik * following test prevents us from completing the qc 1847c6fd2807SJeff Garzik * twice. If we win, the port is frozen and will be 1848c6fd2807SJeff Garzik * cleaned up by ->post_internal_cmd(). 1849c6fd2807SJeff Garzik */ 1850c6fd2807SJeff Garzik if (qc->flags & ATA_QCFLAG_ACTIVE) { 1851c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_TIMEOUT; 1852c6fd2807SJeff Garzik 1853c6fd2807SJeff Garzik if (ap->ops->error_handler) 1854c6fd2807SJeff Garzik ata_port_freeze(ap); 1855c6fd2807SJeff Garzik else 1856c6fd2807SJeff Garzik ata_qc_complete(qc); 1857c6fd2807SJeff Garzik 1858c6fd2807SJeff Garzik if (ata_msg_warn(ap)) 1859c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_WARNING, 1860c6fd2807SJeff Garzik "qc timeout (cmd 0x%x)\n", command); 1861c6fd2807SJeff Garzik } 1862c6fd2807SJeff Garzik 1863c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 1864c6fd2807SJeff Garzik } 1865c6fd2807SJeff Garzik 1866c6fd2807SJeff Garzik /* do post_internal_cmd */ 1867c6fd2807SJeff Garzik if (ap->ops->post_internal_cmd) 1868c6fd2807SJeff Garzik ap->ops->post_internal_cmd(qc); 1869c6fd2807SJeff Garzik 1870a51d644aSTejun Heo /* perform minimal error analysis */ 1871a51d644aSTejun Heo if (qc->flags & ATA_QCFLAG_FAILED) { 1872a51d644aSTejun Heo if (qc->result_tf.command & (ATA_ERR | ATA_DF)) 1873a51d644aSTejun Heo qc->err_mask |= AC_ERR_DEV; 1874a51d644aSTejun Heo 1875a51d644aSTejun Heo if (!qc->err_mask) 1876c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_OTHER; 1877a51d644aSTejun Heo 1878a51d644aSTejun Heo if (qc->err_mask & ~AC_ERR_OTHER) 1879a51d644aSTejun Heo qc->err_mask &= ~AC_ERR_OTHER; 1880c6fd2807SJeff Garzik } 1881c6fd2807SJeff Garzik 1882c6fd2807SJeff Garzik /* finish up */ 1883c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 1884c6fd2807SJeff Garzik 1885c6fd2807SJeff Garzik *tf = qc->result_tf; 1886c6fd2807SJeff Garzik err_mask = qc->err_mask; 1887c6fd2807SJeff Garzik 1888c6fd2807SJeff Garzik ata_qc_free(qc); 18899af5c9c9STejun Heo link->active_tag = preempted_tag; 18909af5c9c9STejun Heo link->sactive = preempted_sactive; 1891c6fd2807SJeff Garzik ap->qc_active = preempted_qc_active; 1892da917d69STejun Heo ap->nr_active_links = preempted_nr_active_links; 1893c6fd2807SJeff Garzik 1894c6fd2807SJeff Garzik /* XXX - Some LLDDs (sata_mv) disable port on command failure. 1895c6fd2807SJeff Garzik * Until those drivers are fixed, we detect the condition 1896c6fd2807SJeff Garzik * here, fail the command with AC_ERR_SYSTEM and reenable the 1897c6fd2807SJeff Garzik * port. 1898c6fd2807SJeff Garzik * 1899c6fd2807SJeff Garzik * Note that this doesn't change any behavior as internal 1900c6fd2807SJeff Garzik * command failure results in disabling the device in the 1901c6fd2807SJeff Garzik * higher layer for LLDDs without new reset/EH callbacks. 1902c6fd2807SJeff Garzik * 1903c6fd2807SJeff Garzik * Kill the following code as soon as those drivers are fixed. 1904c6fd2807SJeff Garzik */ 1905c6fd2807SJeff Garzik if (ap->flags & ATA_FLAG_DISABLED) { 1906c6fd2807SJeff Garzik err_mask |= AC_ERR_SYSTEM; 1907c6fd2807SJeff Garzik ata_port_probe(ap); 1908c6fd2807SJeff Garzik } 1909c6fd2807SJeff Garzik 1910c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 1911c6fd2807SJeff Garzik 1912c6fd2807SJeff Garzik return err_mask; 1913c6fd2807SJeff Garzik } 1914c6fd2807SJeff Garzik 1915c6fd2807SJeff Garzik /** 191633480a0eSTejun Heo * ata_exec_internal - execute libata internal command 19172432697bSTejun Heo * @dev: Device to which the command is sent 19182432697bSTejun Heo * @tf: Taskfile registers for the command and the result 19192432697bSTejun Heo * @cdb: CDB for packet command 19202432697bSTejun Heo * @dma_dir: Data tranfer direction of the command 19212432697bSTejun Heo * @buf: Data buffer of the command 19222432697bSTejun Heo * @buflen: Length of data buffer 19232b789108STejun Heo * @timeout: Timeout in msecs (0 for default) 19242432697bSTejun Heo * 19252432697bSTejun Heo * Wrapper around ata_exec_internal_sg() which takes simple 19262432697bSTejun Heo * buffer instead of sg list. 19272432697bSTejun Heo * 19282432697bSTejun Heo * LOCKING: 19292432697bSTejun Heo * None. Should be called with kernel context, might sleep. 19302432697bSTejun Heo * 19312432697bSTejun Heo * RETURNS: 19322432697bSTejun Heo * Zero on success, AC_ERR_* mask on failure 19332432697bSTejun Heo */ 19342432697bSTejun Heo unsigned ata_exec_internal(struct ata_device *dev, 19352432697bSTejun Heo struct ata_taskfile *tf, const u8 *cdb, 19362b789108STejun Heo int dma_dir, void *buf, unsigned int buflen, 19372b789108STejun Heo unsigned long timeout) 19382432697bSTejun Heo { 193933480a0eSTejun Heo struct scatterlist *psg = NULL, sg; 194033480a0eSTejun Heo unsigned int n_elem = 0; 19412432697bSTejun Heo 194233480a0eSTejun Heo if (dma_dir != DMA_NONE) { 194333480a0eSTejun Heo WARN_ON(!buf); 19442432697bSTejun Heo sg_init_one(&sg, buf, buflen); 194533480a0eSTejun Heo psg = &sg; 194633480a0eSTejun Heo n_elem++; 194733480a0eSTejun Heo } 19482432697bSTejun Heo 19492b789108STejun Heo return ata_exec_internal_sg(dev, tf, cdb, dma_dir, psg, n_elem, 19502b789108STejun Heo timeout); 19512432697bSTejun Heo } 19522432697bSTejun Heo 19532432697bSTejun Heo /** 1954c6fd2807SJeff Garzik * ata_do_simple_cmd - execute simple internal command 1955c6fd2807SJeff Garzik * @dev: Device to which the command is sent 1956c6fd2807SJeff Garzik * @cmd: Opcode to execute 1957c6fd2807SJeff Garzik * 1958c6fd2807SJeff Garzik * Execute a 'simple' command, that only consists of the opcode 1959c6fd2807SJeff Garzik * 'cmd' itself, without filling any other registers 1960c6fd2807SJeff Garzik * 1961c6fd2807SJeff Garzik * LOCKING: 1962c6fd2807SJeff Garzik * Kernel thread context (may sleep). 1963c6fd2807SJeff Garzik * 1964c6fd2807SJeff Garzik * RETURNS: 1965c6fd2807SJeff Garzik * Zero on success, AC_ERR_* mask on failure 1966c6fd2807SJeff Garzik */ 1967c6fd2807SJeff Garzik unsigned int ata_do_simple_cmd(struct ata_device *dev, u8 cmd) 1968c6fd2807SJeff Garzik { 1969c6fd2807SJeff Garzik struct ata_taskfile tf; 1970c6fd2807SJeff Garzik 1971c6fd2807SJeff Garzik ata_tf_init(dev, &tf); 1972c6fd2807SJeff Garzik 1973c6fd2807SJeff Garzik tf.command = cmd; 1974c6fd2807SJeff Garzik tf.flags |= ATA_TFLAG_DEVICE; 1975c6fd2807SJeff Garzik tf.protocol = ATA_PROT_NODATA; 1976c6fd2807SJeff Garzik 19772b789108STejun Heo return ata_exec_internal(dev, &tf, NULL, DMA_NONE, NULL, 0, 0); 1978c6fd2807SJeff Garzik } 1979c6fd2807SJeff Garzik 1980c6fd2807SJeff Garzik /** 1981c6fd2807SJeff Garzik * ata_pio_need_iordy - check if iordy needed 1982c6fd2807SJeff Garzik * @adev: ATA device 1983c6fd2807SJeff Garzik * 1984c6fd2807SJeff Garzik * Check if the current speed of the device requires IORDY. Used 1985c6fd2807SJeff Garzik * by various controllers for chip configuration. 1986c6fd2807SJeff Garzik */ 1987c6fd2807SJeff Garzik 1988c6fd2807SJeff Garzik unsigned int ata_pio_need_iordy(const struct ata_device *adev) 1989c6fd2807SJeff Garzik { 1990432729f0SAlan Cox /* Controller doesn't support IORDY. Probably a pointless check 1991432729f0SAlan Cox as the caller should know this */ 19929af5c9c9STejun Heo if (adev->link->ap->flags & ATA_FLAG_NO_IORDY) 1993c6fd2807SJeff Garzik return 0; 1994432729f0SAlan Cox /* PIO3 and higher it is mandatory */ 1995432729f0SAlan Cox if (adev->pio_mode > XFER_PIO_2) 1996c6fd2807SJeff Garzik return 1; 1997432729f0SAlan Cox /* We turn it on when possible */ 1998432729f0SAlan Cox if (ata_id_has_iordy(adev->id)) 1999432729f0SAlan Cox return 1; 2000432729f0SAlan Cox return 0; 2001432729f0SAlan Cox } 2002c6fd2807SJeff Garzik 2003432729f0SAlan Cox /** 2004432729f0SAlan Cox * ata_pio_mask_no_iordy - Return the non IORDY mask 2005432729f0SAlan Cox * @adev: ATA device 2006432729f0SAlan Cox * 2007432729f0SAlan Cox * Compute the highest mode possible if we are not using iordy. Return 2008432729f0SAlan Cox * -1 if no iordy mode is available. 2009432729f0SAlan Cox */ 2010432729f0SAlan Cox 2011432729f0SAlan Cox static u32 ata_pio_mask_no_iordy(const struct ata_device *adev) 2012432729f0SAlan Cox { 2013c6fd2807SJeff Garzik /* If we have no drive specific rule, then PIO 2 is non IORDY */ 2014c6fd2807SJeff Garzik if (adev->id[ATA_ID_FIELD_VALID] & 2) { /* EIDE */ 2015432729f0SAlan Cox u16 pio = adev->id[ATA_ID_EIDE_PIO]; 2016c6fd2807SJeff Garzik /* Is the speed faster than the drive allows non IORDY ? */ 2017c6fd2807SJeff Garzik if (pio) { 2018c6fd2807SJeff Garzik /* This is cycle times not frequency - watch the logic! */ 2019c6fd2807SJeff Garzik if (pio > 240) /* PIO2 is 240nS per cycle */ 2020432729f0SAlan Cox return 3 << ATA_SHIFT_PIO; 2021432729f0SAlan Cox return 7 << ATA_SHIFT_PIO; 2022c6fd2807SJeff Garzik } 2023c6fd2807SJeff Garzik } 2024432729f0SAlan Cox return 3 << ATA_SHIFT_PIO; 2025c6fd2807SJeff Garzik } 2026c6fd2807SJeff Garzik 2027c6fd2807SJeff Garzik /** 2028c6fd2807SJeff Garzik * ata_dev_read_id - Read ID data from the specified device 2029c6fd2807SJeff Garzik * @dev: target device 2030c6fd2807SJeff Garzik * @p_class: pointer to class of the target device (may be changed) 2031bff04647STejun Heo * @flags: ATA_READID_* flags 2032c6fd2807SJeff Garzik * @id: buffer to read IDENTIFY data into 2033c6fd2807SJeff Garzik * 2034c6fd2807SJeff Garzik * Read ID data from the specified device. ATA_CMD_ID_ATA is 2035c6fd2807SJeff Garzik * performed on ATA devices and ATA_CMD_ID_ATAPI on ATAPI 2036c6fd2807SJeff Garzik * devices. This function also issues ATA_CMD_INIT_DEV_PARAMS 2037c6fd2807SJeff Garzik * for pre-ATA4 drives. 2038c6fd2807SJeff Garzik * 203950a99018SAlan Cox * FIXME: ATA_CMD_ID_ATA is optional for early drives and right 204050a99018SAlan Cox * now we abort if we hit that case. 204150a99018SAlan Cox * 2042c6fd2807SJeff Garzik * LOCKING: 2043c6fd2807SJeff Garzik * Kernel thread context (may sleep) 2044c6fd2807SJeff Garzik * 2045c6fd2807SJeff Garzik * RETURNS: 2046c6fd2807SJeff Garzik * 0 on success, -errno otherwise. 2047c6fd2807SJeff Garzik */ 2048c6fd2807SJeff Garzik int ata_dev_read_id(struct ata_device *dev, unsigned int *p_class, 2049bff04647STejun Heo unsigned int flags, u16 *id) 2050c6fd2807SJeff Garzik { 20519af5c9c9STejun Heo struct ata_port *ap = dev->link->ap; 2052c6fd2807SJeff Garzik unsigned int class = *p_class; 2053c6fd2807SJeff Garzik struct ata_taskfile tf; 2054c6fd2807SJeff Garzik unsigned int err_mask = 0; 2055c6fd2807SJeff Garzik const char *reason; 205654936f8bSTejun Heo int may_fallback = 1, tried_spinup = 0; 2057c6fd2807SJeff Garzik int rc; 2058c6fd2807SJeff Garzik 2059c6fd2807SJeff Garzik if (ata_msg_ctl(ap)) 20607f5e4e8dSHarvey Harrison ata_dev_printk(dev, KERN_DEBUG, "%s: ENTER\n", __func__); 2061c6fd2807SJeff Garzik 2062c6fd2807SJeff Garzik ata_dev_select(ap, dev->devno, 1, 1); /* select device 0/1 */ 2063c6fd2807SJeff Garzik retry: 2064c6fd2807SJeff Garzik ata_tf_init(dev, &tf); 2065c6fd2807SJeff Garzik 2066c6fd2807SJeff Garzik switch (class) { 2067c6fd2807SJeff Garzik case ATA_DEV_ATA: 2068c6fd2807SJeff Garzik tf.command = ATA_CMD_ID_ATA; 2069c6fd2807SJeff Garzik break; 2070c6fd2807SJeff Garzik case ATA_DEV_ATAPI: 2071c6fd2807SJeff Garzik tf.command = ATA_CMD_ID_ATAPI; 2072c6fd2807SJeff Garzik break; 2073c6fd2807SJeff Garzik default: 2074c6fd2807SJeff Garzik rc = -ENODEV; 2075c6fd2807SJeff Garzik reason = "unsupported class"; 2076c6fd2807SJeff Garzik goto err_out; 2077c6fd2807SJeff Garzik } 2078c6fd2807SJeff Garzik 2079c6fd2807SJeff Garzik tf.protocol = ATA_PROT_PIO; 208081afe893STejun Heo 208181afe893STejun Heo /* Some devices choke if TF registers contain garbage. Make 208281afe893STejun Heo * sure those are properly initialized. 208381afe893STejun Heo */ 208481afe893STejun Heo tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; 208581afe893STejun Heo 208681afe893STejun Heo /* Device presence detection is unreliable on some 208781afe893STejun Heo * controllers. Always poll IDENTIFY if available. 208881afe893STejun Heo */ 208981afe893STejun Heo tf.flags |= ATA_TFLAG_POLLING; 2090c6fd2807SJeff Garzik 2091c6fd2807SJeff Garzik err_mask = ata_exec_internal(dev, &tf, NULL, DMA_FROM_DEVICE, 20922b789108STejun Heo id, sizeof(id[0]) * ATA_ID_WORDS, 0); 2093c6fd2807SJeff Garzik if (err_mask) { 2094800b3996STejun Heo if (err_mask & AC_ERR_NODEV_HINT) { 209555a8e2c8STejun Heo DPRINTK("ata%u.%d: NODEV after polling detection\n", 209644877b4eSTejun Heo ap->print_id, dev->devno); 209755a8e2c8STejun Heo return -ENOENT; 209855a8e2c8STejun Heo } 209955a8e2c8STejun Heo 210054936f8bSTejun Heo /* Device or controller might have reported the wrong 210154936f8bSTejun Heo * device class. Give a shot at the other IDENTIFY if 210254936f8bSTejun Heo * the current one is aborted by the device. 210354936f8bSTejun Heo */ 210454936f8bSTejun Heo if (may_fallback && 210554936f8bSTejun Heo (err_mask == AC_ERR_DEV) && (tf.feature & ATA_ABORTED)) { 210654936f8bSTejun Heo may_fallback = 0; 210754936f8bSTejun Heo 210854936f8bSTejun Heo if (class == ATA_DEV_ATA) 210954936f8bSTejun Heo class = ATA_DEV_ATAPI; 211054936f8bSTejun Heo else 211154936f8bSTejun Heo class = ATA_DEV_ATA; 211254936f8bSTejun Heo goto retry; 211354936f8bSTejun Heo } 211454936f8bSTejun Heo 2115c6fd2807SJeff Garzik rc = -EIO; 2116c6fd2807SJeff Garzik reason = "I/O error"; 2117c6fd2807SJeff Garzik goto err_out; 2118c6fd2807SJeff Garzik } 2119c6fd2807SJeff Garzik 212054936f8bSTejun Heo /* Falling back doesn't make sense if ID data was read 212154936f8bSTejun Heo * successfully at least once. 212254936f8bSTejun Heo */ 212354936f8bSTejun Heo may_fallback = 0; 212454936f8bSTejun Heo 2125c6fd2807SJeff Garzik swap_buf_le16(id, ATA_ID_WORDS); 2126c6fd2807SJeff Garzik 2127c6fd2807SJeff Garzik /* sanity check */ 2128c6fd2807SJeff Garzik rc = -EINVAL; 21296070068bSAlan Cox reason = "device reports invalid type"; 21304a3381feSJeff Garzik 21314a3381feSJeff Garzik if (class == ATA_DEV_ATA) { 21324a3381feSJeff Garzik if (!ata_id_is_ata(id) && !ata_id_is_cfa(id)) 21334a3381feSJeff Garzik goto err_out; 21344a3381feSJeff Garzik } else { 21354a3381feSJeff Garzik if (ata_id_is_ata(id)) 2136c6fd2807SJeff Garzik goto err_out; 2137c6fd2807SJeff Garzik } 2138c6fd2807SJeff Garzik 2139169439c2SMark Lord if (!tried_spinup && (id[2] == 0x37c8 || id[2] == 0x738c)) { 2140169439c2SMark Lord tried_spinup = 1; 2141169439c2SMark Lord /* 2142169439c2SMark Lord * Drive powered-up in standby mode, and requires a specific 2143169439c2SMark Lord * SET_FEATURES spin-up subcommand before it will accept 2144169439c2SMark Lord * anything other than the original IDENTIFY command. 2145169439c2SMark Lord */ 2146218f3d30SJeff Garzik err_mask = ata_dev_set_feature(dev, SETFEATURES_SPINUP, 0); 2147fb0582f9SRyan Power if (err_mask && id[2] != 0x738c) { 2148169439c2SMark Lord rc = -EIO; 2149169439c2SMark Lord reason = "SPINUP failed"; 2150169439c2SMark Lord goto err_out; 2151169439c2SMark Lord } 2152169439c2SMark Lord /* 2153169439c2SMark Lord * If the drive initially returned incomplete IDENTIFY info, 2154169439c2SMark Lord * we now must reissue the IDENTIFY command. 2155169439c2SMark Lord */ 2156169439c2SMark Lord if (id[2] == 0x37c8) 2157169439c2SMark Lord goto retry; 2158169439c2SMark Lord } 2159169439c2SMark Lord 2160bff04647STejun Heo if ((flags & ATA_READID_POSTRESET) && class == ATA_DEV_ATA) { 2161c6fd2807SJeff Garzik /* 2162c6fd2807SJeff Garzik * The exact sequence expected by certain pre-ATA4 drives is: 2163c6fd2807SJeff Garzik * SRST RESET 216450a99018SAlan Cox * IDENTIFY (optional in early ATA) 216550a99018SAlan Cox * INITIALIZE DEVICE PARAMETERS (later IDE and ATA) 2166c6fd2807SJeff Garzik * anything else.. 2167c6fd2807SJeff Garzik * Some drives were very specific about that exact sequence. 216850a99018SAlan Cox * 216950a99018SAlan Cox * Note that ATA4 says lba is mandatory so the second check 217050a99018SAlan Cox * shoud never trigger. 2171c6fd2807SJeff Garzik */ 2172c6fd2807SJeff Garzik if (ata_id_major_version(id) < 4 || !ata_id_has_lba(id)) { 2173c6fd2807SJeff Garzik err_mask = ata_dev_init_params(dev, id[3], id[6]); 2174c6fd2807SJeff Garzik if (err_mask) { 2175c6fd2807SJeff Garzik rc = -EIO; 2176c6fd2807SJeff Garzik reason = "INIT_DEV_PARAMS failed"; 2177c6fd2807SJeff Garzik goto err_out; 2178c6fd2807SJeff Garzik } 2179c6fd2807SJeff Garzik 2180c6fd2807SJeff Garzik /* current CHS translation info (id[53-58]) might be 2181c6fd2807SJeff Garzik * changed. reread the identify device info. 2182c6fd2807SJeff Garzik */ 2183bff04647STejun Heo flags &= ~ATA_READID_POSTRESET; 2184c6fd2807SJeff Garzik goto retry; 2185c6fd2807SJeff Garzik } 2186c6fd2807SJeff Garzik } 2187c6fd2807SJeff Garzik 2188c6fd2807SJeff Garzik *p_class = class; 2189c6fd2807SJeff Garzik 2190c6fd2807SJeff Garzik return 0; 2191c6fd2807SJeff Garzik 2192c6fd2807SJeff Garzik err_out: 2193c6fd2807SJeff Garzik if (ata_msg_warn(ap)) 2194c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_WARNING, "failed to IDENTIFY " 2195c6fd2807SJeff Garzik "(%s, err_mask=0x%x)\n", reason, err_mask); 2196c6fd2807SJeff Garzik return rc; 2197c6fd2807SJeff Garzik } 2198c6fd2807SJeff Garzik 2199c6fd2807SJeff Garzik static inline u8 ata_dev_knobble(struct ata_device *dev) 2200c6fd2807SJeff Garzik { 22019af5c9c9STejun Heo struct ata_port *ap = dev->link->ap; 22029af5c9c9STejun Heo return ((ap->cbl == ATA_CBL_SATA) && (!ata_id_is_sata(dev->id))); 2203c6fd2807SJeff Garzik } 2204c6fd2807SJeff Garzik 2205c6fd2807SJeff Garzik static void ata_dev_config_ncq(struct ata_device *dev, 2206c6fd2807SJeff Garzik char *desc, size_t desc_sz) 2207c6fd2807SJeff Garzik { 22089af5c9c9STejun Heo struct ata_port *ap = dev->link->ap; 2209c6fd2807SJeff Garzik int hdepth = 0, ddepth = ata_id_queue_depth(dev->id); 2210c6fd2807SJeff Garzik 2211c6fd2807SJeff Garzik if (!ata_id_has_ncq(dev->id)) { 2212c6fd2807SJeff Garzik desc[0] = '\0'; 2213c6fd2807SJeff Garzik return; 2214c6fd2807SJeff Garzik } 221575683fe7STejun Heo if (dev->horkage & ATA_HORKAGE_NONCQ) { 22166919a0a6SAlan Cox snprintf(desc, desc_sz, "NCQ (not used)"); 22176919a0a6SAlan Cox return; 22186919a0a6SAlan Cox } 2219c6fd2807SJeff Garzik if (ap->flags & ATA_FLAG_NCQ) { 2220cca3974eSJeff Garzik hdepth = min(ap->scsi_host->can_queue, ATA_MAX_QUEUE - 1); 2221c6fd2807SJeff Garzik dev->flags |= ATA_DFLAG_NCQ; 2222c6fd2807SJeff Garzik } 2223c6fd2807SJeff Garzik 2224c6fd2807SJeff Garzik if (hdepth >= ddepth) 2225c6fd2807SJeff Garzik snprintf(desc, desc_sz, "NCQ (depth %d)", ddepth); 2226c6fd2807SJeff Garzik else 2227c6fd2807SJeff Garzik snprintf(desc, desc_sz, "NCQ (depth %d/%d)", hdepth, ddepth); 2228c6fd2807SJeff Garzik } 2229c6fd2807SJeff Garzik 2230c6fd2807SJeff Garzik /** 2231c6fd2807SJeff Garzik * ata_dev_configure - Configure the specified ATA/ATAPI device 2232c6fd2807SJeff Garzik * @dev: Target device to configure 2233c6fd2807SJeff Garzik * 2234c6fd2807SJeff Garzik * Configure @dev according to @dev->id. Generic and low-level 2235c6fd2807SJeff Garzik * driver specific fixups are also applied. 2236c6fd2807SJeff Garzik * 2237c6fd2807SJeff Garzik * LOCKING: 2238c6fd2807SJeff Garzik * Kernel thread context (may sleep) 2239c6fd2807SJeff Garzik * 2240c6fd2807SJeff Garzik * RETURNS: 2241c6fd2807SJeff Garzik * 0 on success, -errno otherwise 2242c6fd2807SJeff Garzik */ 2243efdaedc4STejun Heo int ata_dev_configure(struct ata_device *dev) 2244c6fd2807SJeff Garzik { 22459af5c9c9STejun Heo struct ata_port *ap = dev->link->ap; 22469af5c9c9STejun Heo struct ata_eh_context *ehc = &dev->link->eh_context; 22476746544cSTejun Heo int print_info = ehc->i.flags & ATA_EHI_PRINTINFO; 2248c6fd2807SJeff Garzik const u16 *id = dev->id; 22497dc951aeSTejun Heo unsigned long xfer_mask; 2250b352e57dSAlan Cox char revbuf[7]; /* XYZ-99\0 */ 22513f64f565SEric D. Mudama char fwrevbuf[ATA_ID_FW_REV_LEN+1]; 22523f64f565SEric D. Mudama char modelbuf[ATA_ID_PROD_LEN+1]; 2253c6fd2807SJeff Garzik int rc; 2254c6fd2807SJeff Garzik 2255c6fd2807SJeff Garzik if (!ata_dev_enabled(dev) && ata_msg_info(ap)) { 225644877b4eSTejun Heo ata_dev_printk(dev, KERN_INFO, "%s: ENTER/EXIT -- nodev\n", 22577f5e4e8dSHarvey Harrison __func__); 2258c6fd2807SJeff Garzik return 0; 2259c6fd2807SJeff Garzik } 2260c6fd2807SJeff Garzik 2261c6fd2807SJeff Garzik if (ata_msg_probe(ap)) 22627f5e4e8dSHarvey Harrison ata_dev_printk(dev, KERN_DEBUG, "%s: ENTER\n", __func__); 2263c6fd2807SJeff Garzik 226475683fe7STejun Heo /* set horkage */ 226575683fe7STejun Heo dev->horkage |= ata_dev_blacklisted(dev); 226633267325STejun Heo ata_force_horkage(dev); 226775683fe7STejun Heo 22686746544cSTejun Heo /* let ACPI work its magic */ 22696746544cSTejun Heo rc = ata_acpi_on_devcfg(dev); 22706746544cSTejun Heo if (rc) 22716746544cSTejun Heo return rc; 227208573a86SKristen Carlson Accardi 227305027adcSTejun Heo /* massage HPA, do it early as it might change IDENTIFY data */ 227405027adcSTejun Heo rc = ata_hpa_resize(dev); 227505027adcSTejun Heo if (rc) 227605027adcSTejun Heo return rc; 227705027adcSTejun Heo 2278c6fd2807SJeff Garzik /* print device capabilities */ 2279c6fd2807SJeff Garzik if (ata_msg_probe(ap)) 2280c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_DEBUG, 2281c6fd2807SJeff Garzik "%s: cfg 49:%04x 82:%04x 83:%04x 84:%04x " 2282c6fd2807SJeff Garzik "85:%04x 86:%04x 87:%04x 88:%04x\n", 22837f5e4e8dSHarvey Harrison __func__, 2284c6fd2807SJeff Garzik id[49], id[82], id[83], id[84], 2285c6fd2807SJeff Garzik id[85], id[86], id[87], id[88]); 2286c6fd2807SJeff Garzik 2287c6fd2807SJeff Garzik /* initialize to-be-configured parameters */ 2288c6fd2807SJeff Garzik dev->flags &= ~ATA_DFLAG_CFG_MASK; 2289c6fd2807SJeff Garzik dev->max_sectors = 0; 2290c6fd2807SJeff Garzik dev->cdb_len = 0; 2291c6fd2807SJeff Garzik dev->n_sectors = 0; 2292c6fd2807SJeff Garzik dev->cylinders = 0; 2293c6fd2807SJeff Garzik dev->heads = 0; 2294c6fd2807SJeff Garzik dev->sectors = 0; 2295c6fd2807SJeff Garzik 2296c6fd2807SJeff Garzik /* 2297c6fd2807SJeff Garzik * common ATA, ATAPI feature tests 2298c6fd2807SJeff Garzik */ 2299c6fd2807SJeff Garzik 2300c6fd2807SJeff Garzik /* find max transfer mode; for printk only */ 2301c6fd2807SJeff Garzik xfer_mask = ata_id_xfermask(id); 2302c6fd2807SJeff Garzik 2303c6fd2807SJeff Garzik if (ata_msg_probe(ap)) 2304c6fd2807SJeff Garzik ata_dump_id(id); 2305c6fd2807SJeff Garzik 2306ef143d57SAlbert Lee /* SCSI only uses 4-char revisions, dump full 8 chars from ATA */ 2307ef143d57SAlbert Lee ata_id_c_string(dev->id, fwrevbuf, ATA_ID_FW_REV, 2308ef143d57SAlbert Lee sizeof(fwrevbuf)); 2309ef143d57SAlbert Lee 2310ef143d57SAlbert Lee ata_id_c_string(dev->id, modelbuf, ATA_ID_PROD, 2311ef143d57SAlbert Lee sizeof(modelbuf)); 2312ef143d57SAlbert Lee 2313c6fd2807SJeff Garzik /* ATA-specific feature tests */ 2314c6fd2807SJeff Garzik if (dev->class == ATA_DEV_ATA) { 2315b352e57dSAlan Cox if (ata_id_is_cfa(id)) { 2316b352e57dSAlan Cox if (id[162] & 1) /* CPRM may make this media unusable */ 231744877b4eSTejun Heo ata_dev_printk(dev, KERN_WARNING, 231844877b4eSTejun Heo "supports DRM functions and may " 231944877b4eSTejun Heo "not be fully accessable.\n"); 2320b352e57dSAlan Cox snprintf(revbuf, 7, "CFA"); 2321ae8d4ee7SAlan Cox } else { 2322b352e57dSAlan Cox snprintf(revbuf, 7, "ATA-%d", ata_id_major_version(id)); 2323ae8d4ee7SAlan Cox /* Warn the user if the device has TPM extensions */ 2324ae8d4ee7SAlan Cox if (ata_id_has_tpm(id)) 2325ae8d4ee7SAlan Cox ata_dev_printk(dev, KERN_WARNING, 2326ae8d4ee7SAlan Cox "supports DRM functions and may " 2327ae8d4ee7SAlan Cox "not be fully accessable.\n"); 2328ae8d4ee7SAlan Cox } 2329b352e57dSAlan Cox 2330c6fd2807SJeff Garzik dev->n_sectors = ata_id_n_sectors(id); 2331c6fd2807SJeff Garzik 23323f64f565SEric D. Mudama if (dev->id[59] & 0x100) 23333f64f565SEric D. Mudama dev->multi_count = dev->id[59] & 0xff; 23343f64f565SEric D. Mudama 2335c6fd2807SJeff Garzik if (ata_id_has_lba(id)) { 2336c6fd2807SJeff Garzik const char *lba_desc; 2337c6fd2807SJeff Garzik char ncq_desc[20]; 2338c6fd2807SJeff Garzik 2339c6fd2807SJeff Garzik lba_desc = "LBA"; 2340c6fd2807SJeff Garzik dev->flags |= ATA_DFLAG_LBA; 2341c6fd2807SJeff Garzik if (ata_id_has_lba48(id)) { 2342c6fd2807SJeff Garzik dev->flags |= ATA_DFLAG_LBA48; 2343c6fd2807SJeff Garzik lba_desc = "LBA48"; 23446fc49adbSTejun Heo 23456fc49adbSTejun Heo if (dev->n_sectors >= (1UL << 28) && 23466fc49adbSTejun Heo ata_id_has_flush_ext(id)) 23476fc49adbSTejun Heo dev->flags |= ATA_DFLAG_FLUSH_EXT; 2348c6fd2807SJeff Garzik } 2349c6fd2807SJeff Garzik 2350c6fd2807SJeff Garzik /* config NCQ */ 2351c6fd2807SJeff Garzik ata_dev_config_ncq(dev, ncq_desc, sizeof(ncq_desc)); 2352c6fd2807SJeff Garzik 2353c6fd2807SJeff Garzik /* print device info to dmesg */ 23543f64f565SEric D. Mudama if (ata_msg_drv(ap) && print_info) { 23553f64f565SEric D. Mudama ata_dev_printk(dev, KERN_INFO, 23563f64f565SEric D. Mudama "%s: %s, %s, max %s\n", 23573f64f565SEric D. Mudama revbuf, modelbuf, fwrevbuf, 23583f64f565SEric D. Mudama ata_mode_string(xfer_mask)); 23593f64f565SEric D. Mudama ata_dev_printk(dev, KERN_INFO, 23603f64f565SEric D. Mudama "%Lu sectors, multi %u: %s %s\n", 2361c6fd2807SJeff Garzik (unsigned long long)dev->n_sectors, 23623f64f565SEric D. Mudama dev->multi_count, lba_desc, ncq_desc); 23633f64f565SEric D. Mudama } 2364c6fd2807SJeff Garzik } else { 2365c6fd2807SJeff Garzik /* CHS */ 2366c6fd2807SJeff Garzik 2367c6fd2807SJeff Garzik /* Default translation */ 2368c6fd2807SJeff Garzik dev->cylinders = id[1]; 2369c6fd2807SJeff Garzik dev->heads = id[3]; 2370c6fd2807SJeff Garzik dev->sectors = id[6]; 2371c6fd2807SJeff Garzik 2372c6fd2807SJeff Garzik if (ata_id_current_chs_valid(id)) { 2373c6fd2807SJeff Garzik /* Current CHS translation is valid. */ 2374c6fd2807SJeff Garzik dev->cylinders = id[54]; 2375c6fd2807SJeff Garzik dev->heads = id[55]; 2376c6fd2807SJeff Garzik dev->sectors = id[56]; 2377c6fd2807SJeff Garzik } 2378c6fd2807SJeff Garzik 2379c6fd2807SJeff Garzik /* print device info to dmesg */ 23803f64f565SEric D. Mudama if (ata_msg_drv(ap) && print_info) { 2381c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_INFO, 23823f64f565SEric D. Mudama "%s: %s, %s, max %s\n", 23833f64f565SEric D. Mudama revbuf, modelbuf, fwrevbuf, 23843f64f565SEric D. Mudama ata_mode_string(xfer_mask)); 23853f64f565SEric D. Mudama ata_dev_printk(dev, KERN_INFO, 23863f64f565SEric D. Mudama "%Lu sectors, multi %u, CHS %u/%u/%u\n", 23873f64f565SEric D. Mudama (unsigned long long)dev->n_sectors, 23883f64f565SEric D. Mudama dev->multi_count, dev->cylinders, 23893f64f565SEric D. Mudama dev->heads, dev->sectors); 23903f64f565SEric D. Mudama } 2391c6fd2807SJeff Garzik } 2392c6fd2807SJeff Garzik 2393c6fd2807SJeff Garzik dev->cdb_len = 16; 2394c6fd2807SJeff Garzik } 2395c6fd2807SJeff Garzik 2396c6fd2807SJeff Garzik /* ATAPI-specific feature tests */ 2397c6fd2807SJeff Garzik else if (dev->class == ATA_DEV_ATAPI) { 2398854c73a2STejun Heo const char *cdb_intr_string = ""; 2399854c73a2STejun Heo const char *atapi_an_string = ""; 240091163006STejun Heo const char *dma_dir_string = ""; 24017d77b247STejun Heo u32 sntf; 2402c6fd2807SJeff Garzik 2403c6fd2807SJeff Garzik rc = atapi_cdb_len(id); 2404c6fd2807SJeff Garzik if ((rc < 12) || (rc > ATAPI_CDB_LEN)) { 2405c6fd2807SJeff Garzik if (ata_msg_warn(ap)) 2406c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_WARNING, 2407c6fd2807SJeff Garzik "unsupported CDB len\n"); 2408c6fd2807SJeff Garzik rc = -EINVAL; 2409c6fd2807SJeff Garzik goto err_out_nosup; 2410c6fd2807SJeff Garzik } 2411c6fd2807SJeff Garzik dev->cdb_len = (unsigned int) rc; 2412c6fd2807SJeff Garzik 24137d77b247STejun Heo /* Enable ATAPI AN if both the host and device have 24147d77b247STejun Heo * the support. If PMP is attached, SNTF is required 24157d77b247STejun Heo * to enable ATAPI AN to discern between PHY status 24167d77b247STejun Heo * changed notifications and ATAPI ANs. 24179f45cbd3SKristen Carlson Accardi */ 24187d77b247STejun Heo if ((ap->flags & ATA_FLAG_AN) && ata_id_has_atapi_AN(id) && 24197d77b247STejun Heo (!ap->nr_pmp_links || 24207d77b247STejun Heo sata_scr_read(&ap->link, SCR_NOTIFICATION, &sntf) == 0)) { 2421854c73a2STejun Heo unsigned int err_mask; 2422854c73a2STejun Heo 24239f45cbd3SKristen Carlson Accardi /* issue SET feature command to turn this on */ 2424218f3d30SJeff Garzik err_mask = ata_dev_set_feature(dev, 2425218f3d30SJeff Garzik SETFEATURES_SATA_ENABLE, SATA_AN); 2426854c73a2STejun Heo if (err_mask) 24279f45cbd3SKristen Carlson Accardi ata_dev_printk(dev, KERN_ERR, 2428854c73a2STejun Heo "failed to enable ATAPI AN " 2429854c73a2STejun Heo "(err_mask=0x%x)\n", err_mask); 2430854c73a2STejun Heo else { 24319f45cbd3SKristen Carlson Accardi dev->flags |= ATA_DFLAG_AN; 2432854c73a2STejun Heo atapi_an_string = ", ATAPI AN"; 2433854c73a2STejun Heo } 24349f45cbd3SKristen Carlson Accardi } 24359f45cbd3SKristen Carlson Accardi 2436c6fd2807SJeff Garzik if (ata_id_cdb_intr(dev->id)) { 2437c6fd2807SJeff Garzik dev->flags |= ATA_DFLAG_CDB_INTR; 2438c6fd2807SJeff Garzik cdb_intr_string = ", CDB intr"; 2439c6fd2807SJeff Garzik } 2440c6fd2807SJeff Garzik 244191163006STejun Heo if (atapi_dmadir || atapi_id_dmadir(dev->id)) { 244291163006STejun Heo dev->flags |= ATA_DFLAG_DMADIR; 244391163006STejun Heo dma_dir_string = ", DMADIR"; 244491163006STejun Heo } 244591163006STejun Heo 2446c6fd2807SJeff Garzik /* print device info to dmesg */ 2447c6fd2807SJeff Garzik if (ata_msg_drv(ap) && print_info) 2448ef143d57SAlbert Lee ata_dev_printk(dev, KERN_INFO, 244991163006STejun Heo "ATAPI: %s, %s, max %s%s%s%s\n", 2450ef143d57SAlbert Lee modelbuf, fwrevbuf, 2451c6fd2807SJeff Garzik ata_mode_string(xfer_mask), 245291163006STejun Heo cdb_intr_string, atapi_an_string, 245391163006STejun Heo dma_dir_string); 2454c6fd2807SJeff Garzik } 2455c6fd2807SJeff Garzik 2456914ed354STejun Heo /* determine max_sectors */ 2457914ed354STejun Heo dev->max_sectors = ATA_MAX_SECTORS; 2458914ed354STejun Heo if (dev->flags & ATA_DFLAG_LBA48) 2459914ed354STejun Heo dev->max_sectors = ATA_MAX_SECTORS_LBA48; 2460914ed354STejun Heo 2461ca77329fSKristen Carlson Accardi if (!(dev->horkage & ATA_HORKAGE_IPM)) { 2462ca77329fSKristen Carlson Accardi if (ata_id_has_hipm(dev->id)) 2463ca77329fSKristen Carlson Accardi dev->flags |= ATA_DFLAG_HIPM; 2464ca77329fSKristen Carlson Accardi if (ata_id_has_dipm(dev->id)) 2465ca77329fSKristen Carlson Accardi dev->flags |= ATA_DFLAG_DIPM; 2466ca77329fSKristen Carlson Accardi } 2467ca77329fSKristen Carlson Accardi 2468c5038fc0SAlan Cox /* Limit PATA drive on SATA cable bridge transfers to udma5, 2469c5038fc0SAlan Cox 200 sectors */ 2470c6fd2807SJeff Garzik if (ata_dev_knobble(dev)) { 2471c6fd2807SJeff Garzik if (ata_msg_drv(ap) && print_info) 2472c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_INFO, 2473c6fd2807SJeff Garzik "applying bridge limits\n"); 2474c6fd2807SJeff Garzik dev->udma_mask &= ATA_UDMA5; 2475c6fd2807SJeff Garzik dev->max_sectors = ATA_MAX_SECTORS; 2476c6fd2807SJeff Garzik } 2477c6fd2807SJeff Garzik 2478f8d8e579STony Battersby if ((dev->class == ATA_DEV_ATAPI) && 2479f442cd86SAlbert Lee (atapi_command_packet_set(id) == TYPE_TAPE)) { 2480f8d8e579STony Battersby dev->max_sectors = ATA_MAX_SECTORS_TAPE; 2481f442cd86SAlbert Lee dev->horkage |= ATA_HORKAGE_STUCK_ERR; 2482f442cd86SAlbert Lee } 2483f8d8e579STony Battersby 248475683fe7STejun Heo if (dev->horkage & ATA_HORKAGE_MAX_SEC_128) 248503ec52deSTejun Heo dev->max_sectors = min_t(unsigned int, ATA_MAX_SECTORS_128, 248603ec52deSTejun Heo dev->max_sectors); 248718d6e9d5SAlbert Lee 2488ca77329fSKristen Carlson Accardi if (ata_dev_blacklisted(dev) & ATA_HORKAGE_IPM) { 2489ca77329fSKristen Carlson Accardi dev->horkage |= ATA_HORKAGE_IPM; 2490ca77329fSKristen Carlson Accardi 2491ca77329fSKristen Carlson Accardi /* reset link pm_policy for this port to no pm */ 2492ca77329fSKristen Carlson Accardi ap->pm_policy = MAX_PERFORMANCE; 2493ca77329fSKristen Carlson Accardi } 2494ca77329fSKristen Carlson Accardi 2495c6fd2807SJeff Garzik if (ap->ops->dev_config) 2496cd0d3bbcSAlan ap->ops->dev_config(dev); 2497c6fd2807SJeff Garzik 2498c5038fc0SAlan Cox if (dev->horkage & ATA_HORKAGE_DIAGNOSTIC) { 2499c5038fc0SAlan Cox /* Let the user know. We don't want to disallow opens for 2500c5038fc0SAlan Cox rescue purposes, or in case the vendor is just a blithering 2501c5038fc0SAlan Cox idiot. Do this after the dev_config call as some controllers 2502c5038fc0SAlan Cox with buggy firmware may want to avoid reporting false device 2503c5038fc0SAlan Cox bugs */ 2504c5038fc0SAlan Cox 2505c5038fc0SAlan Cox if (print_info) { 2506c5038fc0SAlan Cox ata_dev_printk(dev, KERN_WARNING, 2507c5038fc0SAlan Cox "Drive reports diagnostics failure. This may indicate a drive\n"); 2508c5038fc0SAlan Cox ata_dev_printk(dev, KERN_WARNING, 2509c5038fc0SAlan Cox "fault or invalid emulation. Contact drive vendor for information.\n"); 2510c5038fc0SAlan Cox } 2511c5038fc0SAlan Cox } 2512c5038fc0SAlan Cox 2513c6fd2807SJeff Garzik if (ata_msg_probe(ap)) 2514c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_DEBUG, "%s: EXIT, drv_stat = 0x%x\n", 25157f5e4e8dSHarvey Harrison __func__, ata_chk_status(ap)); 2516c6fd2807SJeff Garzik return 0; 2517c6fd2807SJeff Garzik 2518c6fd2807SJeff Garzik err_out_nosup: 2519c6fd2807SJeff Garzik if (ata_msg_probe(ap)) 2520c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_DEBUG, 25217f5e4e8dSHarvey Harrison "%s: EXIT, err\n", __func__); 2522c6fd2807SJeff Garzik return rc; 2523c6fd2807SJeff Garzik } 2524c6fd2807SJeff Garzik 2525c6fd2807SJeff Garzik /** 25262e41e8e6SAlan Cox * ata_cable_40wire - return 40 wire cable type 2527be0d18dfSAlan Cox * @ap: port 2528be0d18dfSAlan Cox * 25292e41e8e6SAlan Cox * Helper method for drivers which want to hardwire 40 wire cable 2530be0d18dfSAlan Cox * detection. 2531be0d18dfSAlan Cox */ 2532be0d18dfSAlan Cox 2533be0d18dfSAlan Cox int ata_cable_40wire(struct ata_port *ap) 2534be0d18dfSAlan Cox { 2535be0d18dfSAlan Cox return ATA_CBL_PATA40; 2536be0d18dfSAlan Cox } 2537be0d18dfSAlan Cox 2538be0d18dfSAlan Cox /** 25392e41e8e6SAlan Cox * ata_cable_80wire - return 80 wire cable type 2540be0d18dfSAlan Cox * @ap: port 2541be0d18dfSAlan Cox * 25422e41e8e6SAlan Cox * Helper method for drivers which want to hardwire 80 wire cable 2543be0d18dfSAlan Cox * detection. 2544be0d18dfSAlan Cox */ 2545be0d18dfSAlan Cox 2546be0d18dfSAlan Cox int ata_cable_80wire(struct ata_port *ap) 2547be0d18dfSAlan Cox { 2548be0d18dfSAlan Cox return ATA_CBL_PATA80; 2549be0d18dfSAlan Cox } 2550be0d18dfSAlan Cox 2551be0d18dfSAlan Cox /** 2552be0d18dfSAlan Cox * ata_cable_unknown - return unknown PATA cable. 2553be0d18dfSAlan Cox * @ap: port 2554be0d18dfSAlan Cox * 2555be0d18dfSAlan Cox * Helper method for drivers which have no PATA cable detection. 2556be0d18dfSAlan Cox */ 2557be0d18dfSAlan Cox 2558be0d18dfSAlan Cox int ata_cable_unknown(struct ata_port *ap) 2559be0d18dfSAlan Cox { 2560be0d18dfSAlan Cox return ATA_CBL_PATA_UNK; 2561be0d18dfSAlan Cox } 2562be0d18dfSAlan Cox 2563be0d18dfSAlan Cox /** 2564c88f90c3STejun Heo * ata_cable_ignore - return ignored PATA cable. 2565c88f90c3STejun Heo * @ap: port 2566c88f90c3STejun Heo * 2567c88f90c3STejun Heo * Helper method for drivers which don't use cable type to limit 2568c88f90c3STejun Heo * transfer mode. 2569c88f90c3STejun Heo */ 2570c88f90c3STejun Heo int ata_cable_ignore(struct ata_port *ap) 2571c88f90c3STejun Heo { 2572c88f90c3STejun Heo return ATA_CBL_PATA_IGN; 2573c88f90c3STejun Heo } 2574c88f90c3STejun Heo 2575c88f90c3STejun Heo /** 2576be0d18dfSAlan Cox * ata_cable_sata - return SATA cable type 2577be0d18dfSAlan Cox * @ap: port 2578be0d18dfSAlan Cox * 2579be0d18dfSAlan Cox * Helper method for drivers which have SATA cables 2580be0d18dfSAlan Cox */ 2581be0d18dfSAlan Cox 2582be0d18dfSAlan Cox int ata_cable_sata(struct ata_port *ap) 2583be0d18dfSAlan Cox { 2584be0d18dfSAlan Cox return ATA_CBL_SATA; 2585be0d18dfSAlan Cox } 2586be0d18dfSAlan Cox 2587be0d18dfSAlan Cox /** 2588c6fd2807SJeff Garzik * ata_bus_probe - Reset and probe ATA bus 2589c6fd2807SJeff Garzik * @ap: Bus to probe 2590c6fd2807SJeff Garzik * 2591c6fd2807SJeff Garzik * Master ATA bus probing function. Initiates a hardware-dependent 2592c6fd2807SJeff Garzik * bus reset, then attempts to identify any devices found on 2593c6fd2807SJeff Garzik * the bus. 2594c6fd2807SJeff Garzik * 2595c6fd2807SJeff Garzik * LOCKING: 2596c6fd2807SJeff Garzik * PCI/etc. bus probe sem. 2597c6fd2807SJeff Garzik * 2598c6fd2807SJeff Garzik * RETURNS: 2599c6fd2807SJeff Garzik * Zero on success, negative errno otherwise. 2600c6fd2807SJeff Garzik */ 2601c6fd2807SJeff Garzik 2602c6fd2807SJeff Garzik int ata_bus_probe(struct ata_port *ap) 2603c6fd2807SJeff Garzik { 2604c6fd2807SJeff Garzik unsigned int classes[ATA_MAX_DEVICES]; 2605c6fd2807SJeff Garzik int tries[ATA_MAX_DEVICES]; 2606f58229f8STejun Heo int rc; 2607c6fd2807SJeff Garzik struct ata_device *dev; 2608c6fd2807SJeff Garzik 2609c6fd2807SJeff Garzik ata_port_probe(ap); 2610c6fd2807SJeff Garzik 2611f58229f8STejun Heo ata_link_for_each_dev(dev, &ap->link) 2612f58229f8STejun Heo tries[dev->devno] = ATA_PROBE_MAX_TRIES; 2613c6fd2807SJeff Garzik 2614c6fd2807SJeff Garzik retry: 2615cdeab114STejun Heo ata_link_for_each_dev(dev, &ap->link) { 2616cdeab114STejun Heo /* If we issue an SRST then an ATA drive (not ATAPI) 2617cdeab114STejun Heo * may change configuration and be in PIO0 timing. If 2618cdeab114STejun Heo * we do a hard reset (or are coming from power on) 2619cdeab114STejun Heo * this is true for ATA or ATAPI. Until we've set a 2620cdeab114STejun Heo * suitable controller mode we should not touch the 2621cdeab114STejun Heo * bus as we may be talking too fast. 2622cdeab114STejun Heo */ 2623cdeab114STejun Heo dev->pio_mode = XFER_PIO_0; 2624cdeab114STejun Heo 2625cdeab114STejun Heo /* If the controller has a pio mode setup function 2626cdeab114STejun Heo * then use it to set the chipset to rights. Don't 2627cdeab114STejun Heo * touch the DMA setup as that will be dealt with when 2628cdeab114STejun Heo * configuring devices. 2629cdeab114STejun Heo */ 2630cdeab114STejun Heo if (ap->ops->set_piomode) 2631cdeab114STejun Heo ap->ops->set_piomode(ap, dev); 2632cdeab114STejun Heo } 2633cdeab114STejun Heo 2634c6fd2807SJeff Garzik /* reset and determine device classes */ 2635c6fd2807SJeff Garzik ap->ops->phy_reset(ap); 2636c6fd2807SJeff Garzik 2637f58229f8STejun Heo ata_link_for_each_dev(dev, &ap->link) { 2638c6fd2807SJeff Garzik if (!(ap->flags & ATA_FLAG_DISABLED) && 2639c6fd2807SJeff Garzik dev->class != ATA_DEV_UNKNOWN) 2640c6fd2807SJeff Garzik classes[dev->devno] = dev->class; 2641c6fd2807SJeff Garzik else 2642c6fd2807SJeff Garzik classes[dev->devno] = ATA_DEV_NONE; 2643c6fd2807SJeff Garzik 2644c6fd2807SJeff Garzik dev->class = ATA_DEV_UNKNOWN; 2645c6fd2807SJeff Garzik } 2646c6fd2807SJeff Garzik 2647c6fd2807SJeff Garzik ata_port_probe(ap); 2648c6fd2807SJeff Garzik 2649f31f0cc2SJeff Garzik /* read IDENTIFY page and configure devices. We have to do the identify 2650f31f0cc2SJeff Garzik specific sequence bass-ackwards so that PDIAG- is released by 2651f31f0cc2SJeff Garzik the slave device */ 2652f31f0cc2SJeff Garzik 2653f58229f8STejun Heo ata_link_for_each_dev(dev, &ap->link) { 2654f58229f8STejun Heo if (tries[dev->devno]) 2655f58229f8STejun Heo dev->class = classes[dev->devno]; 2656c6fd2807SJeff Garzik 2657c6fd2807SJeff Garzik if (!ata_dev_enabled(dev)) 2658c6fd2807SJeff Garzik continue; 2659c6fd2807SJeff Garzik 2660bff04647STejun Heo rc = ata_dev_read_id(dev, &dev->class, ATA_READID_POSTRESET, 2661bff04647STejun Heo dev->id); 2662c6fd2807SJeff Garzik if (rc) 2663c6fd2807SJeff Garzik goto fail; 2664f31f0cc2SJeff Garzik } 2665f31f0cc2SJeff Garzik 2666be0d18dfSAlan Cox /* Now ask for the cable type as PDIAG- should have been released */ 2667be0d18dfSAlan Cox if (ap->ops->cable_detect) 2668be0d18dfSAlan Cox ap->cbl = ap->ops->cable_detect(ap); 2669be0d18dfSAlan Cox 2670614fe29bSAlan Cox /* We may have SATA bridge glue hiding here irrespective of the 2671614fe29bSAlan Cox reported cable types and sensed types */ 2672614fe29bSAlan Cox ata_link_for_each_dev(dev, &ap->link) { 2673614fe29bSAlan Cox if (!ata_dev_enabled(dev)) 2674614fe29bSAlan Cox continue; 2675614fe29bSAlan Cox /* SATA drives indicate we have a bridge. We don't know which 2676614fe29bSAlan Cox end of the link the bridge is which is a problem */ 2677614fe29bSAlan Cox if (ata_id_is_sata(dev->id)) 2678614fe29bSAlan Cox ap->cbl = ATA_CBL_SATA; 2679614fe29bSAlan Cox } 2680614fe29bSAlan Cox 2681f31f0cc2SJeff Garzik /* After the identify sequence we can now set up the devices. We do 2682f31f0cc2SJeff Garzik this in the normal order so that the user doesn't get confused */ 2683f31f0cc2SJeff Garzik 2684f58229f8STejun Heo ata_link_for_each_dev(dev, &ap->link) { 2685f31f0cc2SJeff Garzik if (!ata_dev_enabled(dev)) 2686f31f0cc2SJeff Garzik continue; 2687c6fd2807SJeff Garzik 26889af5c9c9STejun Heo ap->link.eh_context.i.flags |= ATA_EHI_PRINTINFO; 2689efdaedc4STejun Heo rc = ata_dev_configure(dev); 26909af5c9c9STejun Heo ap->link.eh_context.i.flags &= ~ATA_EHI_PRINTINFO; 2691c6fd2807SJeff Garzik if (rc) 2692c6fd2807SJeff Garzik goto fail; 2693c6fd2807SJeff Garzik } 2694c6fd2807SJeff Garzik 2695c6fd2807SJeff Garzik /* configure transfer mode */ 26960260731fSTejun Heo rc = ata_set_mode(&ap->link, &dev); 26974ae72a1eSTejun Heo if (rc) 2698c6fd2807SJeff Garzik goto fail; 2699c6fd2807SJeff Garzik 2700f58229f8STejun Heo ata_link_for_each_dev(dev, &ap->link) 2701f58229f8STejun Heo if (ata_dev_enabled(dev)) 2702c6fd2807SJeff Garzik return 0; 2703c6fd2807SJeff Garzik 2704c6fd2807SJeff Garzik /* no device present, disable port */ 2705c6fd2807SJeff Garzik ata_port_disable(ap); 2706c6fd2807SJeff Garzik return -ENODEV; 2707c6fd2807SJeff Garzik 2708c6fd2807SJeff Garzik fail: 27094ae72a1eSTejun Heo tries[dev->devno]--; 27104ae72a1eSTejun Heo 2711c6fd2807SJeff Garzik switch (rc) { 2712c6fd2807SJeff Garzik case -EINVAL: 27134ae72a1eSTejun Heo /* eeek, something went very wrong, give up */ 2714c6fd2807SJeff Garzik tries[dev->devno] = 0; 2715c6fd2807SJeff Garzik break; 27164ae72a1eSTejun Heo 27174ae72a1eSTejun Heo case -ENODEV: 27184ae72a1eSTejun Heo /* give it just one more chance */ 27194ae72a1eSTejun Heo tries[dev->devno] = min(tries[dev->devno], 1); 2720c6fd2807SJeff Garzik case -EIO: 27214ae72a1eSTejun Heo if (tries[dev->devno] == 1) { 27224ae72a1eSTejun Heo /* This is the last chance, better to slow 27234ae72a1eSTejun Heo * down than lose it. 27244ae72a1eSTejun Heo */ 2725936fd732STejun Heo sata_down_spd_limit(&ap->link); 27264ae72a1eSTejun Heo ata_down_xfermask_limit(dev, ATA_DNXFER_PIO); 27274ae72a1eSTejun Heo } 2728c6fd2807SJeff Garzik } 2729c6fd2807SJeff Garzik 27304ae72a1eSTejun Heo if (!tries[dev->devno]) 2731c6fd2807SJeff Garzik ata_dev_disable(dev); 2732c6fd2807SJeff Garzik 2733c6fd2807SJeff Garzik goto retry; 2734c6fd2807SJeff Garzik } 2735c6fd2807SJeff Garzik 2736c6fd2807SJeff Garzik /** 2737c6fd2807SJeff Garzik * ata_port_probe - Mark port as enabled 2738c6fd2807SJeff Garzik * @ap: Port for which we indicate enablement 2739c6fd2807SJeff Garzik * 2740c6fd2807SJeff Garzik * Modify @ap data structure such that the system 2741c6fd2807SJeff Garzik * thinks that the entire port is enabled. 2742c6fd2807SJeff Garzik * 2743cca3974eSJeff Garzik * LOCKING: host lock, or some other form of 2744c6fd2807SJeff Garzik * serialization. 2745c6fd2807SJeff Garzik */ 2746c6fd2807SJeff Garzik 2747c6fd2807SJeff Garzik void ata_port_probe(struct ata_port *ap) 2748c6fd2807SJeff Garzik { 2749c6fd2807SJeff Garzik ap->flags &= ~ATA_FLAG_DISABLED; 2750c6fd2807SJeff Garzik } 2751c6fd2807SJeff Garzik 2752c6fd2807SJeff Garzik /** 2753c6fd2807SJeff Garzik * sata_print_link_status - Print SATA link status 2754936fd732STejun Heo * @link: SATA link to printk link status about 2755c6fd2807SJeff Garzik * 2756c6fd2807SJeff Garzik * This function prints link speed and status of a SATA link. 2757c6fd2807SJeff Garzik * 2758c6fd2807SJeff Garzik * LOCKING: 2759c6fd2807SJeff Garzik * None. 2760c6fd2807SJeff Garzik */ 2761936fd732STejun Heo void sata_print_link_status(struct ata_link *link) 2762c6fd2807SJeff Garzik { 2763c6fd2807SJeff Garzik u32 sstatus, scontrol, tmp; 2764c6fd2807SJeff Garzik 2765936fd732STejun Heo if (sata_scr_read(link, SCR_STATUS, &sstatus)) 2766c6fd2807SJeff Garzik return; 2767936fd732STejun Heo sata_scr_read(link, SCR_CONTROL, &scontrol); 2768c6fd2807SJeff Garzik 2769936fd732STejun Heo if (ata_link_online(link)) { 2770c6fd2807SJeff Garzik tmp = (sstatus >> 4) & 0xf; 2771936fd732STejun Heo ata_link_printk(link, KERN_INFO, 2772c6fd2807SJeff Garzik "SATA link up %s (SStatus %X SControl %X)\n", 2773c6fd2807SJeff Garzik sata_spd_string(tmp), sstatus, scontrol); 2774c6fd2807SJeff Garzik } else { 2775936fd732STejun Heo ata_link_printk(link, KERN_INFO, 2776c6fd2807SJeff Garzik "SATA link down (SStatus %X SControl %X)\n", 2777c6fd2807SJeff Garzik sstatus, scontrol); 2778c6fd2807SJeff Garzik } 2779c6fd2807SJeff Garzik } 2780c6fd2807SJeff Garzik 2781c6fd2807SJeff Garzik /** 2782c6fd2807SJeff Garzik * ata_dev_pair - return other device on cable 2783c6fd2807SJeff Garzik * @adev: device 2784c6fd2807SJeff Garzik * 2785c6fd2807SJeff Garzik * Obtain the other device on the same cable, or if none is 2786c6fd2807SJeff Garzik * present NULL is returned 2787c6fd2807SJeff Garzik */ 2788c6fd2807SJeff Garzik 2789c6fd2807SJeff Garzik struct ata_device *ata_dev_pair(struct ata_device *adev) 2790c6fd2807SJeff Garzik { 27919af5c9c9STejun Heo struct ata_link *link = adev->link; 27929af5c9c9STejun Heo struct ata_device *pair = &link->device[1 - adev->devno]; 2793c6fd2807SJeff Garzik if (!ata_dev_enabled(pair)) 2794c6fd2807SJeff Garzik return NULL; 2795c6fd2807SJeff Garzik return pair; 2796c6fd2807SJeff Garzik } 2797c6fd2807SJeff Garzik 2798c6fd2807SJeff Garzik /** 2799c6fd2807SJeff Garzik * ata_port_disable - Disable port. 2800c6fd2807SJeff Garzik * @ap: Port to be disabled. 2801c6fd2807SJeff Garzik * 2802c6fd2807SJeff Garzik * Modify @ap data structure such that the system 2803c6fd2807SJeff Garzik * thinks that the entire port is disabled, and should 2804c6fd2807SJeff Garzik * never attempt to probe or communicate with devices 2805c6fd2807SJeff Garzik * on this port. 2806c6fd2807SJeff Garzik * 2807cca3974eSJeff Garzik * LOCKING: host lock, or some other form of 2808c6fd2807SJeff Garzik * serialization. 2809c6fd2807SJeff Garzik */ 2810c6fd2807SJeff Garzik 2811c6fd2807SJeff Garzik void ata_port_disable(struct ata_port *ap) 2812c6fd2807SJeff Garzik { 28139af5c9c9STejun Heo ap->link.device[0].class = ATA_DEV_NONE; 28149af5c9c9STejun Heo ap->link.device[1].class = ATA_DEV_NONE; 2815c6fd2807SJeff Garzik ap->flags |= ATA_FLAG_DISABLED; 2816c6fd2807SJeff Garzik } 2817c6fd2807SJeff Garzik 2818c6fd2807SJeff Garzik /** 2819c6fd2807SJeff Garzik * sata_down_spd_limit - adjust SATA spd limit downward 2820936fd732STejun Heo * @link: Link to adjust SATA spd limit for 2821c6fd2807SJeff Garzik * 2822936fd732STejun Heo * Adjust SATA spd limit of @link downward. Note that this 2823c6fd2807SJeff Garzik * function only adjusts the limit. The change must be applied 2824c6fd2807SJeff Garzik * using sata_set_spd(). 2825c6fd2807SJeff Garzik * 2826c6fd2807SJeff Garzik * LOCKING: 2827c6fd2807SJeff Garzik * Inherited from caller. 2828c6fd2807SJeff Garzik * 2829c6fd2807SJeff Garzik * RETURNS: 2830c6fd2807SJeff Garzik * 0 on success, negative errno on failure 2831c6fd2807SJeff Garzik */ 2832936fd732STejun Heo int sata_down_spd_limit(struct ata_link *link) 2833c6fd2807SJeff Garzik { 2834c6fd2807SJeff Garzik u32 sstatus, spd, mask; 2835c6fd2807SJeff Garzik int rc, highbit; 2836c6fd2807SJeff Garzik 2837936fd732STejun Heo if (!sata_scr_valid(link)) 2838008a7896STejun Heo return -EOPNOTSUPP; 2839008a7896STejun Heo 2840008a7896STejun Heo /* If SCR can be read, use it to determine the current SPD. 2841936fd732STejun Heo * If not, use cached value in link->sata_spd. 2842008a7896STejun Heo */ 2843936fd732STejun Heo rc = sata_scr_read(link, SCR_STATUS, &sstatus); 2844008a7896STejun Heo if (rc == 0) 2845008a7896STejun Heo spd = (sstatus >> 4) & 0xf; 2846008a7896STejun Heo else 2847936fd732STejun Heo spd = link->sata_spd; 2848c6fd2807SJeff Garzik 2849936fd732STejun Heo mask = link->sata_spd_limit; 2850c6fd2807SJeff Garzik if (mask <= 1) 2851c6fd2807SJeff Garzik return -EINVAL; 2852008a7896STejun Heo 2853008a7896STejun Heo /* unconditionally mask off the highest bit */ 2854c6fd2807SJeff Garzik highbit = fls(mask) - 1; 2855c6fd2807SJeff Garzik mask &= ~(1 << highbit); 2856c6fd2807SJeff Garzik 2857008a7896STejun Heo /* Mask off all speeds higher than or equal to the current 2858008a7896STejun Heo * one. Force 1.5Gbps if current SPD is not available. 2859008a7896STejun Heo */ 2860008a7896STejun Heo if (spd > 1) 2861008a7896STejun Heo mask &= (1 << (spd - 1)) - 1; 2862008a7896STejun Heo else 2863008a7896STejun Heo mask &= 1; 2864008a7896STejun Heo 2865008a7896STejun Heo /* were we already at the bottom? */ 2866c6fd2807SJeff Garzik if (!mask) 2867c6fd2807SJeff Garzik return -EINVAL; 2868c6fd2807SJeff Garzik 2869936fd732STejun Heo link->sata_spd_limit = mask; 2870c6fd2807SJeff Garzik 2871936fd732STejun Heo ata_link_printk(link, KERN_WARNING, "limiting SATA link speed to %s\n", 2872c6fd2807SJeff Garzik sata_spd_string(fls(mask))); 2873c6fd2807SJeff Garzik 2874c6fd2807SJeff Garzik return 0; 2875c6fd2807SJeff Garzik } 2876c6fd2807SJeff Garzik 2877936fd732STejun Heo static int __sata_set_spd_needed(struct ata_link *link, u32 *scontrol) 2878c6fd2807SJeff Garzik { 28795270222fSTejun Heo struct ata_link *host_link = &link->ap->link; 28805270222fSTejun Heo u32 limit, target, spd; 2881c6fd2807SJeff Garzik 28825270222fSTejun Heo limit = link->sata_spd_limit; 28835270222fSTejun Heo 28845270222fSTejun Heo /* Don't configure downstream link faster than upstream link. 28855270222fSTejun Heo * It doesn't speed up anything and some PMPs choke on such 28865270222fSTejun Heo * configuration. 28875270222fSTejun Heo */ 28885270222fSTejun Heo if (!ata_is_host_link(link) && host_link->sata_spd) 28895270222fSTejun Heo limit &= (1 << host_link->sata_spd) - 1; 28905270222fSTejun Heo 28915270222fSTejun Heo if (limit == UINT_MAX) 28925270222fSTejun Heo target = 0; 2893c6fd2807SJeff Garzik else 28945270222fSTejun Heo target = fls(limit); 2895c6fd2807SJeff Garzik 2896c6fd2807SJeff Garzik spd = (*scontrol >> 4) & 0xf; 28975270222fSTejun Heo *scontrol = (*scontrol & ~0xf0) | ((target & 0xf) << 4); 2898c6fd2807SJeff Garzik 28995270222fSTejun Heo return spd != target; 2900c6fd2807SJeff Garzik } 2901c6fd2807SJeff Garzik 2902c6fd2807SJeff Garzik /** 2903c6fd2807SJeff Garzik * sata_set_spd_needed - is SATA spd configuration needed 2904936fd732STejun Heo * @link: Link in question 2905c6fd2807SJeff Garzik * 2906c6fd2807SJeff Garzik * Test whether the spd limit in SControl matches 2907936fd732STejun Heo * @link->sata_spd_limit. This function is used to determine 2908c6fd2807SJeff Garzik * whether hardreset is necessary to apply SATA spd 2909c6fd2807SJeff Garzik * configuration. 2910c6fd2807SJeff Garzik * 2911c6fd2807SJeff Garzik * LOCKING: 2912c6fd2807SJeff Garzik * Inherited from caller. 2913c6fd2807SJeff Garzik * 2914c6fd2807SJeff Garzik * RETURNS: 2915c6fd2807SJeff Garzik * 1 if SATA spd configuration is needed, 0 otherwise. 2916c6fd2807SJeff Garzik */ 2917936fd732STejun Heo int sata_set_spd_needed(struct ata_link *link) 2918c6fd2807SJeff Garzik { 2919c6fd2807SJeff Garzik u32 scontrol; 2920c6fd2807SJeff Garzik 2921936fd732STejun Heo if (sata_scr_read(link, SCR_CONTROL, &scontrol)) 2922db64bcf3STejun Heo return 1; 2923c6fd2807SJeff Garzik 2924936fd732STejun Heo return __sata_set_spd_needed(link, &scontrol); 2925c6fd2807SJeff Garzik } 2926c6fd2807SJeff Garzik 2927c6fd2807SJeff Garzik /** 2928c6fd2807SJeff Garzik * sata_set_spd - set SATA spd according to spd limit 2929936fd732STejun Heo * @link: Link to set SATA spd for 2930c6fd2807SJeff Garzik * 2931936fd732STejun Heo * Set SATA spd of @link according to sata_spd_limit. 2932c6fd2807SJeff Garzik * 2933c6fd2807SJeff Garzik * LOCKING: 2934c6fd2807SJeff Garzik * Inherited from caller. 2935c6fd2807SJeff Garzik * 2936c6fd2807SJeff Garzik * RETURNS: 2937c6fd2807SJeff Garzik * 0 if spd doesn't need to be changed, 1 if spd has been 2938c6fd2807SJeff Garzik * changed. Negative errno if SCR registers are inaccessible. 2939c6fd2807SJeff Garzik */ 2940936fd732STejun Heo int sata_set_spd(struct ata_link *link) 2941c6fd2807SJeff Garzik { 2942c6fd2807SJeff Garzik u32 scontrol; 2943c6fd2807SJeff Garzik int rc; 2944c6fd2807SJeff Garzik 2945936fd732STejun Heo if ((rc = sata_scr_read(link, SCR_CONTROL, &scontrol))) 2946c6fd2807SJeff Garzik return rc; 2947c6fd2807SJeff Garzik 2948936fd732STejun Heo if (!__sata_set_spd_needed(link, &scontrol)) 2949c6fd2807SJeff Garzik return 0; 2950c6fd2807SJeff Garzik 2951936fd732STejun Heo if ((rc = sata_scr_write(link, SCR_CONTROL, scontrol))) 2952c6fd2807SJeff Garzik return rc; 2953c6fd2807SJeff Garzik 2954c6fd2807SJeff Garzik return 1; 2955c6fd2807SJeff Garzik } 2956c6fd2807SJeff Garzik 2957c6fd2807SJeff Garzik /* 2958c6fd2807SJeff Garzik * This mode timing computation functionality is ported over from 2959c6fd2807SJeff Garzik * drivers/ide/ide-timing.h and was originally written by Vojtech Pavlik 2960c6fd2807SJeff Garzik */ 2961c6fd2807SJeff Garzik /* 2962b352e57dSAlan Cox * PIO 0-4, MWDMA 0-2 and UDMA 0-6 timings (in nanoseconds). 2963c6fd2807SJeff Garzik * These were taken from ATA/ATAPI-6 standard, rev 0a, except 2964b352e57dSAlan Cox * for UDMA6, which is currently supported only by Maxtor drives. 2965b352e57dSAlan Cox * 2966b352e57dSAlan Cox * For PIO 5/6 MWDMA 3/4 see the CFA specification 3.0. 2967c6fd2807SJeff Garzik */ 2968c6fd2807SJeff Garzik 2969c6fd2807SJeff Garzik static const struct ata_timing ata_timing[] = { 297070cd071eSTejun Heo /* { XFER_PIO_SLOW, 120, 290, 240, 960, 290, 240, 960, 0 }, */ 297170cd071eSTejun Heo { XFER_PIO_0, 70, 290, 240, 600, 165, 150, 600, 0 }, 297270cd071eSTejun Heo { XFER_PIO_1, 50, 290, 93, 383, 125, 100, 383, 0 }, 297370cd071eSTejun Heo { XFER_PIO_2, 30, 290, 40, 330, 100, 90, 240, 0 }, 297470cd071eSTejun Heo { XFER_PIO_3, 30, 80, 70, 180, 80, 70, 180, 0 }, 297570cd071eSTejun Heo { XFER_PIO_4, 25, 70, 25, 120, 70, 25, 120, 0 }, 297670cd071eSTejun Heo { XFER_PIO_5, 15, 65, 25, 100, 65, 25, 100, 0 }, 297770cd071eSTejun Heo { XFER_PIO_6, 10, 55, 20, 80, 55, 20, 80, 0 }, 2978c6fd2807SJeff Garzik 297970cd071eSTejun Heo { XFER_SW_DMA_0, 120, 0, 0, 0, 480, 480, 960, 0 }, 298070cd071eSTejun Heo { XFER_SW_DMA_1, 90, 0, 0, 0, 240, 240, 480, 0 }, 298170cd071eSTejun Heo { XFER_SW_DMA_2, 60, 0, 0, 0, 120, 120, 240, 0 }, 2982c6fd2807SJeff Garzik 298370cd071eSTejun Heo { XFER_MW_DMA_0, 60, 0, 0, 0, 215, 215, 480, 0 }, 298470cd071eSTejun Heo { XFER_MW_DMA_1, 45, 0, 0, 0, 80, 50, 150, 0 }, 298570cd071eSTejun Heo { XFER_MW_DMA_2, 25, 0, 0, 0, 70, 25, 120, 0 }, 2986b352e57dSAlan Cox { XFER_MW_DMA_3, 25, 0, 0, 0, 65, 25, 100, 0 }, 298770cd071eSTejun Heo { XFER_MW_DMA_4, 25, 0, 0, 0, 55, 20, 80, 0 }, 2988c6fd2807SJeff Garzik 2989c6fd2807SJeff Garzik /* { XFER_UDMA_SLOW, 0, 0, 0, 0, 0, 0, 0, 150 }, */ 299070cd071eSTejun Heo { XFER_UDMA_0, 0, 0, 0, 0, 0, 0, 0, 120 }, 299170cd071eSTejun Heo { XFER_UDMA_1, 0, 0, 0, 0, 0, 0, 0, 80 }, 299270cd071eSTejun Heo { XFER_UDMA_2, 0, 0, 0, 0, 0, 0, 0, 60 }, 299370cd071eSTejun Heo { XFER_UDMA_3, 0, 0, 0, 0, 0, 0, 0, 45 }, 299470cd071eSTejun Heo { XFER_UDMA_4, 0, 0, 0, 0, 0, 0, 0, 30 }, 299570cd071eSTejun Heo { XFER_UDMA_5, 0, 0, 0, 0, 0, 0, 0, 20 }, 299670cd071eSTejun Heo { XFER_UDMA_6, 0, 0, 0, 0, 0, 0, 0, 15 }, 2997c6fd2807SJeff Garzik 2998c6fd2807SJeff Garzik { 0xFF } 2999c6fd2807SJeff Garzik }; 3000c6fd2807SJeff Garzik 3001c6fd2807SJeff Garzik #define ENOUGH(v, unit) (((v)-1)/(unit)+1) 3002c6fd2807SJeff Garzik #define EZ(v, unit) ((v)?ENOUGH(v, unit):0) 3003c6fd2807SJeff Garzik 3004c6fd2807SJeff Garzik static void ata_timing_quantize(const struct ata_timing *t, struct ata_timing *q, int T, int UT) 3005c6fd2807SJeff Garzik { 3006c6fd2807SJeff Garzik q->setup = EZ(t->setup * 1000, T); 3007c6fd2807SJeff Garzik q->act8b = EZ(t->act8b * 1000, T); 3008c6fd2807SJeff Garzik q->rec8b = EZ(t->rec8b * 1000, T); 3009c6fd2807SJeff Garzik q->cyc8b = EZ(t->cyc8b * 1000, T); 3010c6fd2807SJeff Garzik q->active = EZ(t->active * 1000, T); 3011c6fd2807SJeff Garzik q->recover = EZ(t->recover * 1000, T); 3012c6fd2807SJeff Garzik q->cycle = EZ(t->cycle * 1000, T); 3013c6fd2807SJeff Garzik q->udma = EZ(t->udma * 1000, UT); 3014c6fd2807SJeff Garzik } 3015c6fd2807SJeff Garzik 3016c6fd2807SJeff Garzik void ata_timing_merge(const struct ata_timing *a, const struct ata_timing *b, 3017c6fd2807SJeff Garzik struct ata_timing *m, unsigned int what) 3018c6fd2807SJeff Garzik { 3019c6fd2807SJeff Garzik if (what & ATA_TIMING_SETUP ) m->setup = max(a->setup, b->setup); 3020c6fd2807SJeff Garzik if (what & ATA_TIMING_ACT8B ) m->act8b = max(a->act8b, b->act8b); 3021c6fd2807SJeff Garzik if (what & ATA_TIMING_REC8B ) m->rec8b = max(a->rec8b, b->rec8b); 3022c6fd2807SJeff Garzik if (what & ATA_TIMING_CYC8B ) m->cyc8b = max(a->cyc8b, b->cyc8b); 3023c6fd2807SJeff Garzik if (what & ATA_TIMING_ACTIVE ) m->active = max(a->active, b->active); 3024c6fd2807SJeff Garzik if (what & ATA_TIMING_RECOVER) m->recover = max(a->recover, b->recover); 3025c6fd2807SJeff Garzik if (what & ATA_TIMING_CYCLE ) m->cycle = max(a->cycle, b->cycle); 3026c6fd2807SJeff Garzik if (what & ATA_TIMING_UDMA ) m->udma = max(a->udma, b->udma); 3027c6fd2807SJeff Garzik } 3028c6fd2807SJeff Garzik 30296357357cSTejun Heo const struct ata_timing *ata_timing_find_mode(u8 xfer_mode) 3030c6fd2807SJeff Garzik { 303170cd071eSTejun Heo const struct ata_timing *t = ata_timing; 3032c6fd2807SJeff Garzik 303370cd071eSTejun Heo while (xfer_mode > t->mode) 303470cd071eSTejun Heo t++; 303570cd071eSTejun Heo 303670cd071eSTejun Heo if (xfer_mode == t->mode) 3037c6fd2807SJeff Garzik return t; 303870cd071eSTejun Heo return NULL; 3039c6fd2807SJeff Garzik } 3040c6fd2807SJeff Garzik 3041c6fd2807SJeff Garzik int ata_timing_compute(struct ata_device *adev, unsigned short speed, 3042c6fd2807SJeff Garzik struct ata_timing *t, int T, int UT) 3043c6fd2807SJeff Garzik { 3044c6fd2807SJeff Garzik const struct ata_timing *s; 3045c6fd2807SJeff Garzik struct ata_timing p; 3046c6fd2807SJeff Garzik 3047c6fd2807SJeff Garzik /* 3048c6fd2807SJeff Garzik * Find the mode. 3049c6fd2807SJeff Garzik */ 3050c6fd2807SJeff Garzik 3051c6fd2807SJeff Garzik if (!(s = ata_timing_find_mode(speed))) 3052c6fd2807SJeff Garzik return -EINVAL; 3053c6fd2807SJeff Garzik 3054c6fd2807SJeff Garzik memcpy(t, s, sizeof(*s)); 3055c6fd2807SJeff Garzik 3056c6fd2807SJeff Garzik /* 3057c6fd2807SJeff Garzik * If the drive is an EIDE drive, it can tell us it needs extended 3058c6fd2807SJeff Garzik * PIO/MW_DMA cycle timing. 3059c6fd2807SJeff Garzik */ 3060c6fd2807SJeff Garzik 3061c6fd2807SJeff Garzik if (adev->id[ATA_ID_FIELD_VALID] & 2) { /* EIDE drive */ 3062c6fd2807SJeff Garzik memset(&p, 0, sizeof(p)); 3063c6fd2807SJeff Garzik if (speed >= XFER_PIO_0 && speed <= XFER_SW_DMA_0) { 3064c6fd2807SJeff Garzik if (speed <= XFER_PIO_2) p.cycle = p.cyc8b = adev->id[ATA_ID_EIDE_PIO]; 3065c6fd2807SJeff Garzik else p.cycle = p.cyc8b = adev->id[ATA_ID_EIDE_PIO_IORDY]; 3066c6fd2807SJeff Garzik } else if (speed >= XFER_MW_DMA_0 && speed <= XFER_MW_DMA_2) { 3067c6fd2807SJeff Garzik p.cycle = adev->id[ATA_ID_EIDE_DMA_MIN]; 3068c6fd2807SJeff Garzik } 3069c6fd2807SJeff Garzik ata_timing_merge(&p, t, t, ATA_TIMING_CYCLE | ATA_TIMING_CYC8B); 3070c6fd2807SJeff Garzik } 3071c6fd2807SJeff Garzik 3072c6fd2807SJeff Garzik /* 3073c6fd2807SJeff Garzik * Convert the timing to bus clock counts. 3074c6fd2807SJeff Garzik */ 3075c6fd2807SJeff Garzik 3076c6fd2807SJeff Garzik ata_timing_quantize(t, t, T, UT); 3077c6fd2807SJeff Garzik 3078c6fd2807SJeff Garzik /* 3079c6fd2807SJeff Garzik * Even in DMA/UDMA modes we still use PIO access for IDENTIFY, 3080c6fd2807SJeff Garzik * S.M.A.R.T * and some other commands. We have to ensure that the 3081c6fd2807SJeff Garzik * DMA cycle timing is slower/equal than the fastest PIO timing. 3082c6fd2807SJeff Garzik */ 3083c6fd2807SJeff Garzik 3084fd3367afSAlan if (speed > XFER_PIO_6) { 3085c6fd2807SJeff Garzik ata_timing_compute(adev, adev->pio_mode, &p, T, UT); 3086c6fd2807SJeff Garzik ata_timing_merge(&p, t, t, ATA_TIMING_ALL); 3087c6fd2807SJeff Garzik } 3088c6fd2807SJeff Garzik 3089c6fd2807SJeff Garzik /* 3090c6fd2807SJeff Garzik * Lengthen active & recovery time so that cycle time is correct. 3091c6fd2807SJeff Garzik */ 3092c6fd2807SJeff Garzik 3093c6fd2807SJeff Garzik if (t->act8b + t->rec8b < t->cyc8b) { 3094c6fd2807SJeff Garzik t->act8b += (t->cyc8b - (t->act8b + t->rec8b)) / 2; 3095c6fd2807SJeff Garzik t->rec8b = t->cyc8b - t->act8b; 3096c6fd2807SJeff Garzik } 3097c6fd2807SJeff Garzik 3098c6fd2807SJeff Garzik if (t->active + t->recover < t->cycle) { 3099c6fd2807SJeff Garzik t->active += (t->cycle - (t->active + t->recover)) / 2; 3100c6fd2807SJeff Garzik t->recover = t->cycle - t->active; 3101c6fd2807SJeff Garzik } 31024f701d1eSAlan Cox 31034f701d1eSAlan Cox /* In a few cases quantisation may produce enough errors to 31044f701d1eSAlan Cox leave t->cycle too low for the sum of active and recovery 31054f701d1eSAlan Cox if so we must correct this */ 31064f701d1eSAlan Cox if (t->active + t->recover > t->cycle) 31074f701d1eSAlan Cox t->cycle = t->active + t->recover; 3108c6fd2807SJeff Garzik 3109c6fd2807SJeff Garzik return 0; 3110c6fd2807SJeff Garzik } 3111c6fd2807SJeff Garzik 3112c6fd2807SJeff Garzik /** 3113a0f79b92STejun Heo * ata_timing_cycle2mode - find xfer mode for the specified cycle duration 3114a0f79b92STejun Heo * @xfer_shift: ATA_SHIFT_* value for transfer type to examine. 3115a0f79b92STejun Heo * @cycle: cycle duration in ns 3116a0f79b92STejun Heo * 3117a0f79b92STejun Heo * Return matching xfer mode for @cycle. The returned mode is of 3118a0f79b92STejun Heo * the transfer type specified by @xfer_shift. If @cycle is too 3119a0f79b92STejun Heo * slow for @xfer_shift, 0xff is returned. If @cycle is faster 3120a0f79b92STejun Heo * than the fastest known mode, the fasted mode is returned. 3121a0f79b92STejun Heo * 3122a0f79b92STejun Heo * LOCKING: 3123a0f79b92STejun Heo * None. 3124a0f79b92STejun Heo * 3125a0f79b92STejun Heo * RETURNS: 3126a0f79b92STejun Heo * Matching xfer_mode, 0xff if no match found. 3127a0f79b92STejun Heo */ 3128a0f79b92STejun Heo u8 ata_timing_cycle2mode(unsigned int xfer_shift, int cycle) 3129a0f79b92STejun Heo { 3130a0f79b92STejun Heo u8 base_mode = 0xff, last_mode = 0xff; 3131a0f79b92STejun Heo const struct ata_xfer_ent *ent; 3132a0f79b92STejun Heo const struct ata_timing *t; 3133a0f79b92STejun Heo 3134a0f79b92STejun Heo for (ent = ata_xfer_tbl; ent->shift >= 0; ent++) 3135a0f79b92STejun Heo if (ent->shift == xfer_shift) 3136a0f79b92STejun Heo base_mode = ent->base; 3137a0f79b92STejun Heo 3138a0f79b92STejun Heo for (t = ata_timing_find_mode(base_mode); 3139a0f79b92STejun Heo t && ata_xfer_mode2shift(t->mode) == xfer_shift; t++) { 3140a0f79b92STejun Heo unsigned short this_cycle; 3141a0f79b92STejun Heo 3142a0f79b92STejun Heo switch (xfer_shift) { 3143a0f79b92STejun Heo case ATA_SHIFT_PIO: 3144a0f79b92STejun Heo case ATA_SHIFT_MWDMA: 3145a0f79b92STejun Heo this_cycle = t->cycle; 3146a0f79b92STejun Heo break; 3147a0f79b92STejun Heo case ATA_SHIFT_UDMA: 3148a0f79b92STejun Heo this_cycle = t->udma; 3149a0f79b92STejun Heo break; 3150a0f79b92STejun Heo default: 3151a0f79b92STejun Heo return 0xff; 3152a0f79b92STejun Heo } 3153a0f79b92STejun Heo 3154a0f79b92STejun Heo if (cycle > this_cycle) 3155a0f79b92STejun Heo break; 3156a0f79b92STejun Heo 3157a0f79b92STejun Heo last_mode = t->mode; 3158a0f79b92STejun Heo } 3159a0f79b92STejun Heo 3160a0f79b92STejun Heo return last_mode; 3161a0f79b92STejun Heo } 3162a0f79b92STejun Heo 3163a0f79b92STejun Heo /** 3164c6fd2807SJeff Garzik * ata_down_xfermask_limit - adjust dev xfer masks downward 3165c6fd2807SJeff Garzik * @dev: Device to adjust xfer masks 3166458337dbSTejun Heo * @sel: ATA_DNXFER_* selector 3167c6fd2807SJeff Garzik * 3168c6fd2807SJeff Garzik * Adjust xfer masks of @dev downward. Note that this function 3169c6fd2807SJeff Garzik * does not apply the change. Invoking ata_set_mode() afterwards 3170c6fd2807SJeff Garzik * will apply the limit. 3171c6fd2807SJeff Garzik * 3172c6fd2807SJeff Garzik * LOCKING: 3173c6fd2807SJeff Garzik * Inherited from caller. 3174c6fd2807SJeff Garzik * 3175c6fd2807SJeff Garzik * RETURNS: 3176c6fd2807SJeff Garzik * 0 on success, negative errno on failure 3177c6fd2807SJeff Garzik */ 3178458337dbSTejun Heo int ata_down_xfermask_limit(struct ata_device *dev, unsigned int sel) 3179c6fd2807SJeff Garzik { 3180458337dbSTejun Heo char buf[32]; 31817dc951aeSTejun Heo unsigned long orig_mask, xfer_mask; 31827dc951aeSTejun Heo unsigned long pio_mask, mwdma_mask, udma_mask; 3183458337dbSTejun Heo int quiet, highbit; 3184c6fd2807SJeff Garzik 3185458337dbSTejun Heo quiet = !!(sel & ATA_DNXFER_QUIET); 3186458337dbSTejun Heo sel &= ~ATA_DNXFER_QUIET; 3187458337dbSTejun Heo 3188458337dbSTejun Heo xfer_mask = orig_mask = ata_pack_xfermask(dev->pio_mask, 3189458337dbSTejun Heo dev->mwdma_mask, 3190c6fd2807SJeff Garzik dev->udma_mask); 3191458337dbSTejun Heo ata_unpack_xfermask(xfer_mask, &pio_mask, &mwdma_mask, &udma_mask); 3192c6fd2807SJeff Garzik 3193458337dbSTejun Heo switch (sel) { 3194458337dbSTejun Heo case ATA_DNXFER_PIO: 3195458337dbSTejun Heo highbit = fls(pio_mask) - 1; 3196458337dbSTejun Heo pio_mask &= ~(1 << highbit); 3197458337dbSTejun Heo break; 3198458337dbSTejun Heo 3199458337dbSTejun Heo case ATA_DNXFER_DMA: 3200458337dbSTejun Heo if (udma_mask) { 3201458337dbSTejun Heo highbit = fls(udma_mask) - 1; 3202458337dbSTejun Heo udma_mask &= ~(1 << highbit); 3203458337dbSTejun Heo if (!udma_mask) 3204458337dbSTejun Heo return -ENOENT; 3205458337dbSTejun Heo } else if (mwdma_mask) { 3206458337dbSTejun Heo highbit = fls(mwdma_mask) - 1; 3207458337dbSTejun Heo mwdma_mask &= ~(1 << highbit); 3208458337dbSTejun Heo if (!mwdma_mask) 3209458337dbSTejun Heo return -ENOENT; 3210458337dbSTejun Heo } 3211458337dbSTejun Heo break; 3212458337dbSTejun Heo 3213458337dbSTejun Heo case ATA_DNXFER_40C: 3214458337dbSTejun Heo udma_mask &= ATA_UDMA_MASK_40C; 3215458337dbSTejun Heo break; 3216458337dbSTejun Heo 3217458337dbSTejun Heo case ATA_DNXFER_FORCE_PIO0: 3218458337dbSTejun Heo pio_mask &= 1; 3219458337dbSTejun Heo case ATA_DNXFER_FORCE_PIO: 3220458337dbSTejun Heo mwdma_mask = 0; 3221458337dbSTejun Heo udma_mask = 0; 3222458337dbSTejun Heo break; 3223458337dbSTejun Heo 3224458337dbSTejun Heo default: 3225458337dbSTejun Heo BUG(); 3226458337dbSTejun Heo } 3227458337dbSTejun Heo 3228458337dbSTejun Heo xfer_mask &= ata_pack_xfermask(pio_mask, mwdma_mask, udma_mask); 3229458337dbSTejun Heo 3230458337dbSTejun Heo if (!(xfer_mask & ATA_MASK_PIO) || xfer_mask == orig_mask) 3231458337dbSTejun Heo return -ENOENT; 3232458337dbSTejun Heo 3233458337dbSTejun Heo if (!quiet) { 3234458337dbSTejun Heo if (xfer_mask & (ATA_MASK_MWDMA | ATA_MASK_UDMA)) 3235458337dbSTejun Heo snprintf(buf, sizeof(buf), "%s:%s", 3236458337dbSTejun Heo ata_mode_string(xfer_mask), 3237458337dbSTejun Heo ata_mode_string(xfer_mask & ATA_MASK_PIO)); 3238458337dbSTejun Heo else 3239458337dbSTejun Heo snprintf(buf, sizeof(buf), "%s", 3240458337dbSTejun Heo ata_mode_string(xfer_mask)); 3241458337dbSTejun Heo 3242458337dbSTejun Heo ata_dev_printk(dev, KERN_WARNING, 3243458337dbSTejun Heo "limiting speed to %s\n", buf); 3244458337dbSTejun Heo } 3245c6fd2807SJeff Garzik 3246c6fd2807SJeff Garzik ata_unpack_xfermask(xfer_mask, &dev->pio_mask, &dev->mwdma_mask, 3247c6fd2807SJeff Garzik &dev->udma_mask); 3248c6fd2807SJeff Garzik 3249c6fd2807SJeff Garzik return 0; 3250c6fd2807SJeff Garzik } 3251c6fd2807SJeff Garzik 3252c6fd2807SJeff Garzik static int ata_dev_set_mode(struct ata_device *dev) 3253c6fd2807SJeff Garzik { 32549af5c9c9STejun Heo struct ata_eh_context *ehc = &dev->link->eh_context; 32554055dee7STejun Heo const char *dev_err_whine = ""; 32564055dee7STejun Heo int ign_dev_err = 0; 3257c6fd2807SJeff Garzik unsigned int err_mask; 3258c6fd2807SJeff Garzik int rc; 3259c6fd2807SJeff Garzik 3260c6fd2807SJeff Garzik dev->flags &= ~ATA_DFLAG_PIO; 3261c6fd2807SJeff Garzik if (dev->xfer_shift == ATA_SHIFT_PIO) 3262c6fd2807SJeff Garzik dev->flags |= ATA_DFLAG_PIO; 3263c6fd2807SJeff Garzik 3264c6fd2807SJeff Garzik err_mask = ata_dev_set_xfermode(dev); 32652dcb407eSJeff Garzik 32664055dee7STejun Heo if (err_mask & ~AC_ERR_DEV) 32674055dee7STejun Heo goto fail; 32682dcb407eSJeff Garzik 32694055dee7STejun Heo /* revalidate */ 3270baa1e78aSTejun Heo ehc->i.flags |= ATA_EHI_POST_SETMODE; 3271422c9daaSTejun Heo rc = ata_dev_revalidate(dev, ATA_DEV_UNKNOWN, 0); 3272baa1e78aSTejun Heo ehc->i.flags &= ~ATA_EHI_POST_SETMODE; 3273c6fd2807SJeff Garzik if (rc) 3274c6fd2807SJeff Garzik return rc; 3275c6fd2807SJeff Garzik 32764055dee7STejun Heo /* Old CFA may refuse this command, which is just fine */ 32774055dee7STejun Heo if (dev->xfer_shift == ATA_SHIFT_PIO && ata_id_is_cfa(dev->id)) 32784055dee7STejun Heo ign_dev_err = 1; 32794055dee7STejun Heo 32804055dee7STejun Heo /* Some very old devices and some bad newer ones fail any kind of 32814055dee7STejun Heo SET_XFERMODE request but support PIO0-2 timings and no IORDY */ 32824055dee7STejun Heo if (dev->xfer_shift == ATA_SHIFT_PIO && !ata_id_has_iordy(dev->id) && 32834055dee7STejun Heo dev->pio_mode <= XFER_PIO_2) 32844055dee7STejun Heo ign_dev_err = 1; 32854055dee7STejun Heo 32864055dee7STejun Heo /* Early MWDMA devices do DMA but don't allow DMA mode setting. 32874055dee7STejun Heo Don't fail an MWDMA0 set IFF the device indicates it is in MWDMA0 */ 32884055dee7STejun Heo if (dev->xfer_shift == ATA_SHIFT_MWDMA && 32894055dee7STejun Heo dev->dma_mode == XFER_MW_DMA_0 && 32904055dee7STejun Heo (dev->id[63] >> 8) & 1) 32914055dee7STejun Heo ign_dev_err = 1; 32924055dee7STejun Heo 32934055dee7STejun Heo /* if the device is actually configured correctly, ignore dev err */ 32944055dee7STejun Heo if (dev->xfer_mode == ata_xfer_mask2mode(ata_id_xfermask(dev->id))) 32954055dee7STejun Heo ign_dev_err = 1; 32964055dee7STejun Heo 32974055dee7STejun Heo if (err_mask & AC_ERR_DEV) { 32984055dee7STejun Heo if (!ign_dev_err) 32994055dee7STejun Heo goto fail; 33004055dee7STejun Heo else 33014055dee7STejun Heo dev_err_whine = " (device error ignored)"; 33024055dee7STejun Heo } 33034055dee7STejun Heo 3304c6fd2807SJeff Garzik DPRINTK("xfer_shift=%u, xfer_mode=0x%x\n", 3305c6fd2807SJeff Garzik dev->xfer_shift, (int)dev->xfer_mode); 3306c6fd2807SJeff Garzik 33074055dee7STejun Heo ata_dev_printk(dev, KERN_INFO, "configured for %s%s\n", 33084055dee7STejun Heo ata_mode_string(ata_xfer_mode2mask(dev->xfer_mode)), 33094055dee7STejun Heo dev_err_whine); 33104055dee7STejun Heo 3311c6fd2807SJeff Garzik return 0; 33124055dee7STejun Heo 33134055dee7STejun Heo fail: 33144055dee7STejun Heo ata_dev_printk(dev, KERN_ERR, "failed to set xfermode " 33154055dee7STejun Heo "(err_mask=0x%x)\n", err_mask); 33164055dee7STejun Heo return -EIO; 3317c6fd2807SJeff Garzik } 3318c6fd2807SJeff Garzik 3319c6fd2807SJeff Garzik /** 332004351821SAlan * ata_do_set_mode - Program timings and issue SET FEATURES - XFER 33210260731fSTejun Heo * @link: link on which timings will be programmed 33221967b7ffSJoe Perches * @r_failed_dev: out parameter for failed device 3323c6fd2807SJeff Garzik * 332404351821SAlan * Standard implementation of the function used to tune and set 332504351821SAlan * ATA device disk transfer mode (PIO3, UDMA6, etc.). If 332604351821SAlan * ata_dev_set_mode() fails, pointer to the failing device is 3327c6fd2807SJeff Garzik * returned in @r_failed_dev. 3328c6fd2807SJeff Garzik * 3329c6fd2807SJeff Garzik * LOCKING: 3330c6fd2807SJeff Garzik * PCI/etc. bus probe sem. 3331c6fd2807SJeff Garzik * 3332c6fd2807SJeff Garzik * RETURNS: 3333c6fd2807SJeff Garzik * 0 on success, negative errno otherwise 3334c6fd2807SJeff Garzik */ 333504351821SAlan 33360260731fSTejun Heo int ata_do_set_mode(struct ata_link *link, struct ata_device **r_failed_dev) 3337c6fd2807SJeff Garzik { 33380260731fSTejun Heo struct ata_port *ap = link->ap; 3339c6fd2807SJeff Garzik struct ata_device *dev; 3340f58229f8STejun Heo int rc = 0, used_dma = 0, found = 0; 3341c6fd2807SJeff Garzik 3342c6fd2807SJeff Garzik /* step 1: calculate xfer_mask */ 3343f58229f8STejun Heo ata_link_for_each_dev(dev, link) { 33447dc951aeSTejun Heo unsigned long pio_mask, dma_mask; 3345b3a70601SAlan Cox unsigned int mode_mask; 3346c6fd2807SJeff Garzik 3347c6fd2807SJeff Garzik if (!ata_dev_enabled(dev)) 3348c6fd2807SJeff Garzik continue; 3349c6fd2807SJeff Garzik 3350b3a70601SAlan Cox mode_mask = ATA_DMA_MASK_ATA; 3351b3a70601SAlan Cox if (dev->class == ATA_DEV_ATAPI) 3352b3a70601SAlan Cox mode_mask = ATA_DMA_MASK_ATAPI; 3353b3a70601SAlan Cox else if (ata_id_is_cfa(dev->id)) 3354b3a70601SAlan Cox mode_mask = ATA_DMA_MASK_CFA; 3355b3a70601SAlan Cox 3356c6fd2807SJeff Garzik ata_dev_xfermask(dev); 335733267325STejun Heo ata_force_xfermask(dev); 3358c6fd2807SJeff Garzik 3359c6fd2807SJeff Garzik pio_mask = ata_pack_xfermask(dev->pio_mask, 0, 0); 3360c6fd2807SJeff Garzik dma_mask = ata_pack_xfermask(0, dev->mwdma_mask, dev->udma_mask); 3361b3a70601SAlan Cox 3362b3a70601SAlan Cox if (libata_dma_mask & mode_mask) 3363b3a70601SAlan Cox dma_mask = ata_pack_xfermask(0, dev->mwdma_mask, dev->udma_mask); 3364b3a70601SAlan Cox else 3365b3a70601SAlan Cox dma_mask = 0; 3366b3a70601SAlan Cox 3367c6fd2807SJeff Garzik dev->pio_mode = ata_xfer_mask2mode(pio_mask); 3368c6fd2807SJeff Garzik dev->dma_mode = ata_xfer_mask2mode(dma_mask); 3369c6fd2807SJeff Garzik 3370c6fd2807SJeff Garzik found = 1; 337170cd071eSTejun Heo if (dev->dma_mode != 0xff) 3372c6fd2807SJeff Garzik used_dma = 1; 3373c6fd2807SJeff Garzik } 3374c6fd2807SJeff Garzik if (!found) 3375c6fd2807SJeff Garzik goto out; 3376c6fd2807SJeff Garzik 3377c6fd2807SJeff Garzik /* step 2: always set host PIO timings */ 3378f58229f8STejun Heo ata_link_for_each_dev(dev, link) { 3379c6fd2807SJeff Garzik if (!ata_dev_enabled(dev)) 3380c6fd2807SJeff Garzik continue; 3381c6fd2807SJeff Garzik 338270cd071eSTejun Heo if (dev->pio_mode == 0xff) { 3383c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_WARNING, "no PIO support\n"); 3384c6fd2807SJeff Garzik rc = -EINVAL; 3385c6fd2807SJeff Garzik goto out; 3386c6fd2807SJeff Garzik } 3387c6fd2807SJeff Garzik 3388c6fd2807SJeff Garzik dev->xfer_mode = dev->pio_mode; 3389c6fd2807SJeff Garzik dev->xfer_shift = ATA_SHIFT_PIO; 3390c6fd2807SJeff Garzik if (ap->ops->set_piomode) 3391c6fd2807SJeff Garzik ap->ops->set_piomode(ap, dev); 3392c6fd2807SJeff Garzik } 3393c6fd2807SJeff Garzik 3394c6fd2807SJeff Garzik /* step 3: set host DMA timings */ 3395f58229f8STejun Heo ata_link_for_each_dev(dev, link) { 339670cd071eSTejun Heo if (!ata_dev_enabled(dev) || dev->dma_mode == 0xff) 3397c6fd2807SJeff Garzik continue; 3398c6fd2807SJeff Garzik 3399c6fd2807SJeff Garzik dev->xfer_mode = dev->dma_mode; 3400c6fd2807SJeff Garzik dev->xfer_shift = ata_xfer_mode2shift(dev->dma_mode); 3401c6fd2807SJeff Garzik if (ap->ops->set_dmamode) 3402c6fd2807SJeff Garzik ap->ops->set_dmamode(ap, dev); 3403c6fd2807SJeff Garzik } 3404c6fd2807SJeff Garzik 3405c6fd2807SJeff Garzik /* step 4: update devices' xfer mode */ 3406f58229f8STejun Heo ata_link_for_each_dev(dev, link) { 340718d90debSAlan /* don't update suspended devices' xfer mode */ 34089666f400STejun Heo if (!ata_dev_enabled(dev)) 3409c6fd2807SJeff Garzik continue; 3410c6fd2807SJeff Garzik 3411c6fd2807SJeff Garzik rc = ata_dev_set_mode(dev); 3412c6fd2807SJeff Garzik if (rc) 3413c6fd2807SJeff Garzik goto out; 3414c6fd2807SJeff Garzik } 3415c6fd2807SJeff Garzik 3416c6fd2807SJeff Garzik /* Record simplex status. If we selected DMA then the other 3417c6fd2807SJeff Garzik * host channels are not permitted to do so. 3418c6fd2807SJeff Garzik */ 3419cca3974eSJeff Garzik if (used_dma && (ap->host->flags & ATA_HOST_SIMPLEX)) 3420032af1ceSAlan ap->host->simplex_claimed = ap; 3421c6fd2807SJeff Garzik 3422c6fd2807SJeff Garzik out: 3423c6fd2807SJeff Garzik if (rc) 3424c6fd2807SJeff Garzik *r_failed_dev = dev; 3425c6fd2807SJeff Garzik return rc; 3426c6fd2807SJeff Garzik } 3427c6fd2807SJeff Garzik 3428c6fd2807SJeff Garzik /** 3429c6fd2807SJeff Garzik * ata_tf_to_host - issue ATA taskfile to host controller 3430c6fd2807SJeff Garzik * @ap: port to which command is being issued 3431c6fd2807SJeff Garzik * @tf: ATA taskfile register set 3432c6fd2807SJeff Garzik * 3433c6fd2807SJeff Garzik * Issues ATA taskfile register set to ATA host controller, 3434c6fd2807SJeff Garzik * with proper synchronization with interrupt handler and 3435c6fd2807SJeff Garzik * other threads. 3436c6fd2807SJeff Garzik * 3437c6fd2807SJeff Garzik * LOCKING: 3438cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 3439c6fd2807SJeff Garzik */ 3440c6fd2807SJeff Garzik 3441c6fd2807SJeff Garzik static inline void ata_tf_to_host(struct ata_port *ap, 3442c6fd2807SJeff Garzik const struct ata_taskfile *tf) 3443c6fd2807SJeff Garzik { 3444c6fd2807SJeff Garzik ap->ops->tf_load(ap, tf); 3445c6fd2807SJeff Garzik ap->ops->exec_command(ap, tf); 3446c6fd2807SJeff Garzik } 3447c6fd2807SJeff Garzik 3448c6fd2807SJeff Garzik /** 3449c6fd2807SJeff Garzik * ata_busy_sleep - sleep until BSY clears, or timeout 3450c6fd2807SJeff Garzik * @ap: port containing status register to be polled 3451c6fd2807SJeff Garzik * @tmout_pat: impatience timeout 3452c6fd2807SJeff Garzik * @tmout: overall timeout 3453c6fd2807SJeff Garzik * 3454c6fd2807SJeff Garzik * Sleep until ATA Status register bit BSY clears, 3455c6fd2807SJeff Garzik * or a timeout occurs. 3456c6fd2807SJeff Garzik * 3457d1adc1bbSTejun Heo * LOCKING: 3458d1adc1bbSTejun Heo * Kernel thread context (may sleep). 3459d1adc1bbSTejun Heo * 3460d1adc1bbSTejun Heo * RETURNS: 3461d1adc1bbSTejun Heo * 0 on success, -errno otherwise. 3462c6fd2807SJeff Garzik */ 3463d1adc1bbSTejun Heo int ata_busy_sleep(struct ata_port *ap, 3464c6fd2807SJeff Garzik unsigned long tmout_pat, unsigned long tmout) 3465c6fd2807SJeff Garzik { 3466c6fd2807SJeff Garzik unsigned long timer_start, timeout; 3467c6fd2807SJeff Garzik u8 status; 3468c6fd2807SJeff Garzik 3469c6fd2807SJeff Garzik status = ata_busy_wait(ap, ATA_BUSY, 300); 3470c6fd2807SJeff Garzik timer_start = jiffies; 3471c6fd2807SJeff Garzik timeout = timer_start + tmout_pat; 3472d1adc1bbSTejun Heo while (status != 0xff && (status & ATA_BUSY) && 3473d1adc1bbSTejun Heo time_before(jiffies, timeout)) { 3474c6fd2807SJeff Garzik msleep(50); 3475c6fd2807SJeff Garzik status = ata_busy_wait(ap, ATA_BUSY, 3); 3476c6fd2807SJeff Garzik } 3477c6fd2807SJeff Garzik 3478d1adc1bbSTejun Heo if (status != 0xff && (status & ATA_BUSY)) 3479c6fd2807SJeff Garzik ata_port_printk(ap, KERN_WARNING, 348035aa7a43SJeff Garzik "port is slow to respond, please be patient " 348135aa7a43SJeff Garzik "(Status 0x%x)\n", status); 3482c6fd2807SJeff Garzik 3483c6fd2807SJeff Garzik timeout = timer_start + tmout; 3484d1adc1bbSTejun Heo while (status != 0xff && (status & ATA_BUSY) && 3485d1adc1bbSTejun Heo time_before(jiffies, timeout)) { 3486c6fd2807SJeff Garzik msleep(50); 3487c6fd2807SJeff Garzik status = ata_chk_status(ap); 3488c6fd2807SJeff Garzik } 3489c6fd2807SJeff Garzik 3490d1adc1bbSTejun Heo if (status == 0xff) 3491d1adc1bbSTejun Heo return -ENODEV; 3492d1adc1bbSTejun Heo 3493c6fd2807SJeff Garzik if (status & ATA_BUSY) { 3494c6fd2807SJeff Garzik ata_port_printk(ap, KERN_ERR, "port failed to respond " 349535aa7a43SJeff Garzik "(%lu secs, Status 0x%x)\n", 349635aa7a43SJeff Garzik tmout / HZ, status); 3497d1adc1bbSTejun Heo return -EBUSY; 3498c6fd2807SJeff Garzik } 3499c6fd2807SJeff Garzik 3500c6fd2807SJeff Garzik return 0; 3501c6fd2807SJeff Garzik } 3502c6fd2807SJeff Garzik 3503d4b2bab4STejun Heo /** 350488ff6eafSTejun Heo * ata_wait_after_reset - wait before checking status after reset 350588ff6eafSTejun Heo * @ap: port containing status register to be polled 350688ff6eafSTejun Heo * @deadline: deadline jiffies for the operation 350788ff6eafSTejun Heo * 350888ff6eafSTejun Heo * After reset, we need to pause a while before reading status. 350988ff6eafSTejun Heo * Also, certain combination of controller and device report 0xff 351088ff6eafSTejun Heo * for some duration (e.g. until SATA PHY is up and running) 351188ff6eafSTejun Heo * which is interpreted as empty port in ATA world. This 351288ff6eafSTejun Heo * function also waits for such devices to get out of 0xff 351388ff6eafSTejun Heo * status. 351488ff6eafSTejun Heo * 351588ff6eafSTejun Heo * LOCKING: 351688ff6eafSTejun Heo * Kernel thread context (may sleep). 351788ff6eafSTejun Heo */ 351888ff6eafSTejun Heo void ata_wait_after_reset(struct ata_port *ap, unsigned long deadline) 351988ff6eafSTejun Heo { 352088ff6eafSTejun Heo unsigned long until = jiffies + ATA_TMOUT_FF_WAIT; 352188ff6eafSTejun Heo 352288ff6eafSTejun Heo if (time_before(until, deadline)) 352388ff6eafSTejun Heo deadline = until; 352488ff6eafSTejun Heo 352588ff6eafSTejun Heo /* Spec mandates ">= 2ms" before checking status. We wait 352688ff6eafSTejun Heo * 150ms, because that was the magic delay used for ATAPI 352788ff6eafSTejun Heo * devices in Hale Landis's ATADRVR, for the period of time 352888ff6eafSTejun Heo * between when the ATA command register is written, and then 352988ff6eafSTejun Heo * status is checked. Because waiting for "a while" before 353088ff6eafSTejun Heo * checking status is fine, post SRST, we perform this magic 353188ff6eafSTejun Heo * delay here as well. 353288ff6eafSTejun Heo * 353388ff6eafSTejun Heo * Old drivers/ide uses the 2mS rule and then waits for ready. 353488ff6eafSTejun Heo */ 353588ff6eafSTejun Heo msleep(150); 353688ff6eafSTejun Heo 353788ff6eafSTejun Heo /* Wait for 0xff to clear. Some SATA devices take a long time 353888ff6eafSTejun Heo * to clear 0xff after reset. For example, HHD424020F7SV00 353988ff6eafSTejun Heo * iVDR needs >= 800ms while. Quantum GoVault needs even more 354088ff6eafSTejun Heo * than that. 35411974e201STejun Heo * 35421974e201STejun Heo * Note that some PATA controllers (pata_ali) explode if 35431974e201STejun Heo * status register is read more than once when there's no 35441974e201STejun Heo * device attached. 354588ff6eafSTejun Heo */ 35461974e201STejun Heo if (ap->flags & ATA_FLAG_SATA) { 354788ff6eafSTejun Heo while (1) { 354888ff6eafSTejun Heo u8 status = ata_chk_status(ap); 354988ff6eafSTejun Heo 355088ff6eafSTejun Heo if (status != 0xff || time_after(jiffies, deadline)) 355188ff6eafSTejun Heo return; 355288ff6eafSTejun Heo 355388ff6eafSTejun Heo msleep(50); 355488ff6eafSTejun Heo } 355588ff6eafSTejun Heo } 35561974e201STejun Heo } 355788ff6eafSTejun Heo 355888ff6eafSTejun Heo /** 3559d4b2bab4STejun Heo * ata_wait_ready - sleep until BSY clears, or timeout 3560d4b2bab4STejun Heo * @ap: port containing status register to be polled 3561d4b2bab4STejun Heo * @deadline: deadline jiffies for the operation 3562d4b2bab4STejun Heo * 3563d4b2bab4STejun Heo * Sleep until ATA Status register bit BSY clears, or timeout 3564d4b2bab4STejun Heo * occurs. 3565d4b2bab4STejun Heo * 3566d4b2bab4STejun Heo * LOCKING: 3567d4b2bab4STejun Heo * Kernel thread context (may sleep). 3568d4b2bab4STejun Heo * 3569d4b2bab4STejun Heo * RETURNS: 3570d4b2bab4STejun Heo * 0 on success, -errno otherwise. 3571d4b2bab4STejun Heo */ 3572d4b2bab4STejun Heo int ata_wait_ready(struct ata_port *ap, unsigned long deadline) 3573d4b2bab4STejun Heo { 3574d4b2bab4STejun Heo unsigned long start = jiffies; 3575d4b2bab4STejun Heo int warned = 0; 3576d4b2bab4STejun Heo 3577d4b2bab4STejun Heo while (1) { 3578d4b2bab4STejun Heo u8 status = ata_chk_status(ap); 3579d4b2bab4STejun Heo unsigned long now = jiffies; 3580d4b2bab4STejun Heo 3581d4b2bab4STejun Heo if (!(status & ATA_BUSY)) 3582d4b2bab4STejun Heo return 0; 3583936fd732STejun Heo if (!ata_link_online(&ap->link) && status == 0xff) 3584d4b2bab4STejun Heo return -ENODEV; 3585d4b2bab4STejun Heo if (time_after(now, deadline)) 3586d4b2bab4STejun Heo return -EBUSY; 3587d4b2bab4STejun Heo 3588d4b2bab4STejun Heo if (!warned && time_after(now, start + 5 * HZ) && 3589d4b2bab4STejun Heo (deadline - now > 3 * HZ)) { 3590d4b2bab4STejun Heo ata_port_printk(ap, KERN_WARNING, 3591d4b2bab4STejun Heo "port is slow to respond, please be patient " 3592d4b2bab4STejun Heo "(Status 0x%x)\n", status); 3593d4b2bab4STejun Heo warned = 1; 3594d4b2bab4STejun Heo } 3595d4b2bab4STejun Heo 3596d4b2bab4STejun Heo msleep(50); 3597d4b2bab4STejun Heo } 3598d4b2bab4STejun Heo } 3599d4b2bab4STejun Heo 3600d4b2bab4STejun Heo static int ata_bus_post_reset(struct ata_port *ap, unsigned int devmask, 3601d4b2bab4STejun Heo unsigned long deadline) 3602c6fd2807SJeff Garzik { 3603c6fd2807SJeff Garzik struct ata_ioports *ioaddr = &ap->ioaddr; 3604c6fd2807SJeff Garzik unsigned int dev0 = devmask & (1 << 0); 3605c6fd2807SJeff Garzik unsigned int dev1 = devmask & (1 << 1); 36069b89391cSTejun Heo int rc, ret = 0; 3607c6fd2807SJeff Garzik 3608c6fd2807SJeff Garzik /* if device 0 was found in ata_devchk, wait for its 3609c6fd2807SJeff Garzik * BSY bit to clear 3610c6fd2807SJeff Garzik */ 3611d4b2bab4STejun Heo if (dev0) { 3612d4b2bab4STejun Heo rc = ata_wait_ready(ap, deadline); 36139b89391cSTejun Heo if (rc) { 36149b89391cSTejun Heo if (rc != -ENODEV) 3615d4b2bab4STejun Heo return rc; 36169b89391cSTejun Heo ret = rc; 36179b89391cSTejun Heo } 3618d4b2bab4STejun Heo } 3619c6fd2807SJeff Garzik 3620e141d999STejun Heo /* if device 1 was found in ata_devchk, wait for register 3621e141d999STejun Heo * access briefly, then wait for BSY to clear. 3622c6fd2807SJeff Garzik */ 3623e141d999STejun Heo if (dev1) { 3624e141d999STejun Heo int i; 3625c6fd2807SJeff Garzik 3626c6fd2807SJeff Garzik ap->ops->dev_select(ap, 1); 3627e141d999STejun Heo 3628e141d999STejun Heo /* Wait for register access. Some ATAPI devices fail 3629e141d999STejun Heo * to set nsect/lbal after reset, so don't waste too 3630e141d999STejun Heo * much time on it. We're gonna wait for !BSY anyway. 3631e141d999STejun Heo */ 3632e141d999STejun Heo for (i = 0; i < 2; i++) { 3633e141d999STejun Heo u8 nsect, lbal; 3634e141d999STejun Heo 36350d5ff566STejun Heo nsect = ioread8(ioaddr->nsect_addr); 36360d5ff566STejun Heo lbal = ioread8(ioaddr->lbal_addr); 3637c6fd2807SJeff Garzik if ((nsect == 1) && (lbal == 1)) 3638c6fd2807SJeff Garzik break; 3639c6fd2807SJeff Garzik msleep(50); /* give drive a breather */ 3640c6fd2807SJeff Garzik } 3641e141d999STejun Heo 3642d4b2bab4STejun Heo rc = ata_wait_ready(ap, deadline); 36439b89391cSTejun Heo if (rc) { 36449b89391cSTejun Heo if (rc != -ENODEV) 3645d4b2bab4STejun Heo return rc; 36469b89391cSTejun Heo ret = rc; 36479b89391cSTejun Heo } 3648d4b2bab4STejun Heo } 3649c6fd2807SJeff Garzik 3650c6fd2807SJeff Garzik /* is all this really necessary? */ 3651c6fd2807SJeff Garzik ap->ops->dev_select(ap, 0); 3652c6fd2807SJeff Garzik if (dev1) 3653c6fd2807SJeff Garzik ap->ops->dev_select(ap, 1); 3654c6fd2807SJeff Garzik if (dev0) 3655c6fd2807SJeff Garzik ap->ops->dev_select(ap, 0); 3656d4b2bab4STejun Heo 36579b89391cSTejun Heo return ret; 3658c6fd2807SJeff Garzik } 3659c6fd2807SJeff Garzik 3660d4b2bab4STejun Heo static int ata_bus_softreset(struct ata_port *ap, unsigned int devmask, 3661d4b2bab4STejun Heo unsigned long deadline) 3662c6fd2807SJeff Garzik { 3663c6fd2807SJeff Garzik struct ata_ioports *ioaddr = &ap->ioaddr; 3664c6fd2807SJeff Garzik 366544877b4eSTejun Heo DPRINTK("ata%u: bus reset via SRST\n", ap->print_id); 3666c6fd2807SJeff Garzik 3667c6fd2807SJeff Garzik /* software reset. causes dev0 to be selected */ 36680d5ff566STejun Heo iowrite8(ap->ctl, ioaddr->ctl_addr); 3669c6fd2807SJeff Garzik udelay(20); /* FIXME: flush */ 36700d5ff566STejun Heo iowrite8(ap->ctl | ATA_SRST, ioaddr->ctl_addr); 3671c6fd2807SJeff Garzik udelay(20); /* FIXME: flush */ 36720d5ff566STejun Heo iowrite8(ap->ctl, ioaddr->ctl_addr); 3673c6fd2807SJeff Garzik 367488ff6eafSTejun Heo /* wait a while before checking status */ 367588ff6eafSTejun Heo ata_wait_after_reset(ap, deadline); 3676c6fd2807SJeff Garzik 3677c6fd2807SJeff Garzik /* Before we perform post reset processing we want to see if 3678c6fd2807SJeff Garzik * the bus shows 0xFF because the odd clown forgets the D7 3679c6fd2807SJeff Garzik * pulldown resistor. 3680c6fd2807SJeff Garzik */ 3681150981b0SAlan Cox if (ata_chk_status(ap) == 0xFF) 36829b89391cSTejun Heo return -ENODEV; 3683c6fd2807SJeff Garzik 3684d4b2bab4STejun Heo return ata_bus_post_reset(ap, devmask, deadline); 3685c6fd2807SJeff Garzik } 3686c6fd2807SJeff Garzik 3687c6fd2807SJeff Garzik /** 3688c6fd2807SJeff Garzik * ata_bus_reset - reset host port and associated ATA channel 3689c6fd2807SJeff Garzik * @ap: port to reset 3690c6fd2807SJeff Garzik * 3691c6fd2807SJeff Garzik * This is typically the first time we actually start issuing 3692c6fd2807SJeff Garzik * commands to the ATA channel. We wait for BSY to clear, then 3693c6fd2807SJeff Garzik * issue EXECUTE DEVICE DIAGNOSTIC command, polling for its 3694c6fd2807SJeff Garzik * result. Determine what devices, if any, are on the channel 3695c6fd2807SJeff Garzik * by looking at the device 0/1 error register. Look at the signature 3696c6fd2807SJeff Garzik * stored in each device's taskfile registers, to determine if 3697c6fd2807SJeff Garzik * the device is ATA or ATAPI. 3698c6fd2807SJeff Garzik * 3699c6fd2807SJeff Garzik * LOCKING: 3700c6fd2807SJeff Garzik * PCI/etc. bus probe sem. 3701cca3974eSJeff Garzik * Obtains host lock. 3702c6fd2807SJeff Garzik * 3703c6fd2807SJeff Garzik * SIDE EFFECTS: 3704c6fd2807SJeff Garzik * Sets ATA_FLAG_DISABLED if bus reset fails. 3705c6fd2807SJeff Garzik */ 3706c6fd2807SJeff Garzik 3707c6fd2807SJeff Garzik void ata_bus_reset(struct ata_port *ap) 3708c6fd2807SJeff Garzik { 37099af5c9c9STejun Heo struct ata_device *device = ap->link.device; 3710c6fd2807SJeff Garzik struct ata_ioports *ioaddr = &ap->ioaddr; 3711c6fd2807SJeff Garzik unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; 3712c6fd2807SJeff Garzik u8 err; 3713c6fd2807SJeff Garzik unsigned int dev0, dev1 = 0, devmask = 0; 37149b89391cSTejun Heo int rc; 3715c6fd2807SJeff Garzik 371644877b4eSTejun Heo DPRINTK("ENTER, host %u, port %u\n", ap->print_id, ap->port_no); 3717c6fd2807SJeff Garzik 3718c6fd2807SJeff Garzik /* determine if device 0/1 are present */ 3719c6fd2807SJeff Garzik if (ap->flags & ATA_FLAG_SATA_RESET) 3720c6fd2807SJeff Garzik dev0 = 1; 3721c6fd2807SJeff Garzik else { 3722c6fd2807SJeff Garzik dev0 = ata_devchk(ap, 0); 3723c6fd2807SJeff Garzik if (slave_possible) 3724c6fd2807SJeff Garzik dev1 = ata_devchk(ap, 1); 3725c6fd2807SJeff Garzik } 3726c6fd2807SJeff Garzik 3727c6fd2807SJeff Garzik if (dev0) 3728c6fd2807SJeff Garzik devmask |= (1 << 0); 3729c6fd2807SJeff Garzik if (dev1) 3730c6fd2807SJeff Garzik devmask |= (1 << 1); 3731c6fd2807SJeff Garzik 3732c6fd2807SJeff Garzik /* select device 0 again */ 3733c6fd2807SJeff Garzik ap->ops->dev_select(ap, 0); 3734c6fd2807SJeff Garzik 3735c6fd2807SJeff Garzik /* issue bus reset */ 37369b89391cSTejun Heo if (ap->flags & ATA_FLAG_SRST) { 37379b89391cSTejun Heo rc = ata_bus_softreset(ap, devmask, jiffies + 40 * HZ); 37389b89391cSTejun Heo if (rc && rc != -ENODEV) 3739c6fd2807SJeff Garzik goto err_out; 37409b89391cSTejun Heo } 3741c6fd2807SJeff Garzik 3742c6fd2807SJeff Garzik /* 3743c6fd2807SJeff Garzik * determine by signature whether we have ATA or ATAPI devices 3744c6fd2807SJeff Garzik */ 37453f19859eSTejun Heo device[0].class = ata_dev_try_classify(&device[0], dev0, &err); 3746c6fd2807SJeff Garzik if ((slave_possible) && (err != 0x81)) 37473f19859eSTejun Heo device[1].class = ata_dev_try_classify(&device[1], dev1, &err); 3748c6fd2807SJeff Garzik 3749c6fd2807SJeff Garzik /* is double-select really necessary? */ 37509af5c9c9STejun Heo if (device[1].class != ATA_DEV_NONE) 3751c6fd2807SJeff Garzik ap->ops->dev_select(ap, 1); 37529af5c9c9STejun Heo if (device[0].class != ATA_DEV_NONE) 3753c6fd2807SJeff Garzik ap->ops->dev_select(ap, 0); 3754c6fd2807SJeff Garzik 3755c6fd2807SJeff Garzik /* if no devices were detected, disable this port */ 37569af5c9c9STejun Heo if ((device[0].class == ATA_DEV_NONE) && 37579af5c9c9STejun Heo (device[1].class == ATA_DEV_NONE)) 3758c6fd2807SJeff Garzik goto err_out; 3759c6fd2807SJeff Garzik 3760c6fd2807SJeff Garzik if (ap->flags & (ATA_FLAG_SATA_RESET | ATA_FLAG_SRST)) { 3761c6fd2807SJeff Garzik /* set up device control for ATA_FLAG_SATA_RESET */ 37620d5ff566STejun Heo iowrite8(ap->ctl, ioaddr->ctl_addr); 3763c6fd2807SJeff Garzik } 3764c6fd2807SJeff Garzik 3765c6fd2807SJeff Garzik DPRINTK("EXIT\n"); 3766c6fd2807SJeff Garzik return; 3767c6fd2807SJeff Garzik 3768c6fd2807SJeff Garzik err_out: 3769c6fd2807SJeff Garzik ata_port_printk(ap, KERN_ERR, "disabling port\n"); 3770ac8869d5SJeff Garzik ata_port_disable(ap); 3771c6fd2807SJeff Garzik 3772c6fd2807SJeff Garzik DPRINTK("EXIT\n"); 3773c6fd2807SJeff Garzik } 3774c6fd2807SJeff Garzik 3775c6fd2807SJeff Garzik /** 3776936fd732STejun Heo * sata_link_debounce - debounce SATA phy status 3777936fd732STejun Heo * @link: ATA link to debounce SATA phy status for 3778c6fd2807SJeff Garzik * @params: timing parameters { interval, duratinon, timeout } in msec 3779d4b2bab4STejun Heo * @deadline: deadline jiffies for the operation 3780c6fd2807SJeff Garzik * 3781936fd732STejun Heo * Make sure SStatus of @link reaches stable state, determined by 3782c6fd2807SJeff Garzik * holding the same value where DET is not 1 for @duration polled 3783c6fd2807SJeff Garzik * every @interval, before @timeout. Timeout constraints the 3784d4b2bab4STejun Heo * beginning of the stable state. Because DET gets stuck at 1 on 3785d4b2bab4STejun Heo * some controllers after hot unplugging, this functions waits 3786c6fd2807SJeff Garzik * until timeout then returns 0 if DET is stable at 1. 3787c6fd2807SJeff Garzik * 3788d4b2bab4STejun Heo * @timeout is further limited by @deadline. The sooner of the 3789d4b2bab4STejun Heo * two is used. 3790d4b2bab4STejun Heo * 3791c6fd2807SJeff Garzik * LOCKING: 3792c6fd2807SJeff Garzik * Kernel thread context (may sleep) 3793c6fd2807SJeff Garzik * 3794c6fd2807SJeff Garzik * RETURNS: 3795c6fd2807SJeff Garzik * 0 on success, -errno on failure. 3796c6fd2807SJeff Garzik */ 3797936fd732STejun Heo int sata_link_debounce(struct ata_link *link, const unsigned long *params, 3798d4b2bab4STejun Heo unsigned long deadline) 3799c6fd2807SJeff Garzik { 3800c6fd2807SJeff Garzik unsigned long interval_msec = params[0]; 3801d4b2bab4STejun Heo unsigned long duration = msecs_to_jiffies(params[1]); 3802d4b2bab4STejun Heo unsigned long last_jiffies, t; 3803c6fd2807SJeff Garzik u32 last, cur; 3804c6fd2807SJeff Garzik int rc; 3805c6fd2807SJeff Garzik 3806d4b2bab4STejun Heo t = jiffies + msecs_to_jiffies(params[2]); 3807d4b2bab4STejun Heo if (time_before(t, deadline)) 3808d4b2bab4STejun Heo deadline = t; 3809d4b2bab4STejun Heo 3810936fd732STejun Heo if ((rc = sata_scr_read(link, SCR_STATUS, &cur))) 3811c6fd2807SJeff Garzik return rc; 3812c6fd2807SJeff Garzik cur &= 0xf; 3813c6fd2807SJeff Garzik 3814c6fd2807SJeff Garzik last = cur; 3815c6fd2807SJeff Garzik last_jiffies = jiffies; 3816c6fd2807SJeff Garzik 3817c6fd2807SJeff Garzik while (1) { 3818c6fd2807SJeff Garzik msleep(interval_msec); 3819936fd732STejun Heo if ((rc = sata_scr_read(link, SCR_STATUS, &cur))) 3820c6fd2807SJeff Garzik return rc; 3821c6fd2807SJeff Garzik cur &= 0xf; 3822c6fd2807SJeff Garzik 3823c6fd2807SJeff Garzik /* DET stable? */ 3824c6fd2807SJeff Garzik if (cur == last) { 3825d4b2bab4STejun Heo if (cur == 1 && time_before(jiffies, deadline)) 3826c6fd2807SJeff Garzik continue; 3827c6fd2807SJeff Garzik if (time_after(jiffies, last_jiffies + duration)) 3828c6fd2807SJeff Garzik return 0; 3829c6fd2807SJeff Garzik continue; 3830c6fd2807SJeff Garzik } 3831c6fd2807SJeff Garzik 3832c6fd2807SJeff Garzik /* unstable, start over */ 3833c6fd2807SJeff Garzik last = cur; 3834c6fd2807SJeff Garzik last_jiffies = jiffies; 3835c6fd2807SJeff Garzik 3836f1545154STejun Heo /* Check deadline. If debouncing failed, return 3837f1545154STejun Heo * -EPIPE to tell upper layer to lower link speed. 3838f1545154STejun Heo */ 3839d4b2bab4STejun Heo if (time_after(jiffies, deadline)) 3840f1545154STejun Heo return -EPIPE; 3841c6fd2807SJeff Garzik } 3842c6fd2807SJeff Garzik } 3843c6fd2807SJeff Garzik 3844c6fd2807SJeff Garzik /** 3845936fd732STejun Heo * sata_link_resume - resume SATA link 3846936fd732STejun Heo * @link: ATA link to resume SATA 3847c6fd2807SJeff Garzik * @params: timing parameters { interval, duratinon, timeout } in msec 3848d4b2bab4STejun Heo * @deadline: deadline jiffies for the operation 3849c6fd2807SJeff Garzik * 3850936fd732STejun Heo * Resume SATA phy @link and debounce it. 3851c6fd2807SJeff Garzik * 3852c6fd2807SJeff Garzik * LOCKING: 3853c6fd2807SJeff Garzik * Kernel thread context (may sleep) 3854c6fd2807SJeff Garzik * 3855c6fd2807SJeff Garzik * RETURNS: 3856c6fd2807SJeff Garzik * 0 on success, -errno on failure. 3857c6fd2807SJeff Garzik */ 3858936fd732STejun Heo int sata_link_resume(struct ata_link *link, const unsigned long *params, 3859d4b2bab4STejun Heo unsigned long deadline) 3860c6fd2807SJeff Garzik { 3861c6fd2807SJeff Garzik u32 scontrol; 3862c6fd2807SJeff Garzik int rc; 3863c6fd2807SJeff Garzik 3864936fd732STejun Heo if ((rc = sata_scr_read(link, SCR_CONTROL, &scontrol))) 3865c6fd2807SJeff Garzik return rc; 3866c6fd2807SJeff Garzik 3867c6fd2807SJeff Garzik scontrol = (scontrol & 0x0f0) | 0x300; 3868c6fd2807SJeff Garzik 3869936fd732STejun Heo if ((rc = sata_scr_write(link, SCR_CONTROL, scontrol))) 3870c6fd2807SJeff Garzik return rc; 3871c6fd2807SJeff Garzik 3872c6fd2807SJeff Garzik /* Some PHYs react badly if SStatus is pounded immediately 3873c6fd2807SJeff Garzik * after resuming. Delay 200ms before debouncing. 3874c6fd2807SJeff Garzik */ 3875c6fd2807SJeff Garzik msleep(200); 3876c6fd2807SJeff Garzik 3877936fd732STejun Heo return sata_link_debounce(link, params, deadline); 3878c6fd2807SJeff Garzik } 3879c6fd2807SJeff Garzik 3880c6fd2807SJeff Garzik /** 3881c6fd2807SJeff Garzik * ata_std_prereset - prepare for reset 3882cc0680a5STejun Heo * @link: ATA link to be reset 3883d4b2bab4STejun Heo * @deadline: deadline jiffies for the operation 3884c6fd2807SJeff Garzik * 3885cc0680a5STejun Heo * @link is about to be reset. Initialize it. Failure from 3886b8cffc6aSTejun Heo * prereset makes libata abort whole reset sequence and give up 3887b8cffc6aSTejun Heo * that port, so prereset should be best-effort. It does its 3888b8cffc6aSTejun Heo * best to prepare for reset sequence but if things go wrong, it 3889b8cffc6aSTejun Heo * should just whine, not fail. 3890c6fd2807SJeff Garzik * 3891c6fd2807SJeff Garzik * LOCKING: 3892c6fd2807SJeff Garzik * Kernel thread context (may sleep) 3893c6fd2807SJeff Garzik * 3894c6fd2807SJeff Garzik * RETURNS: 3895c6fd2807SJeff Garzik * 0 on success, -errno otherwise. 3896c6fd2807SJeff Garzik */ 3897cc0680a5STejun Heo int ata_std_prereset(struct ata_link *link, unsigned long deadline) 3898c6fd2807SJeff Garzik { 3899cc0680a5STejun Heo struct ata_port *ap = link->ap; 3900936fd732STejun Heo struct ata_eh_context *ehc = &link->eh_context; 3901c6fd2807SJeff Garzik const unsigned long *timing = sata_ehc_deb_timing(ehc); 3902c6fd2807SJeff Garzik int rc; 3903c6fd2807SJeff Garzik 390431daabdaSTejun Heo /* handle link resume */ 3905c6fd2807SJeff Garzik if ((ehc->i.flags & ATA_EHI_RESUME_LINK) && 39060c88758bSTejun Heo (link->flags & ATA_LFLAG_HRST_TO_RESUME)) 3907c6fd2807SJeff Garzik ehc->i.action |= ATA_EH_HARDRESET; 3908c6fd2807SJeff Garzik 3909633273a3STejun Heo /* Some PMPs don't work with only SRST, force hardreset if PMP 3910633273a3STejun Heo * is supported. 3911633273a3STejun Heo */ 3912633273a3STejun Heo if (ap->flags & ATA_FLAG_PMP) 3913633273a3STejun Heo ehc->i.action |= ATA_EH_HARDRESET; 3914633273a3STejun Heo 3915c6fd2807SJeff Garzik /* if we're about to do hardreset, nothing more to do */ 3916c6fd2807SJeff Garzik if (ehc->i.action & ATA_EH_HARDRESET) 3917c6fd2807SJeff Garzik return 0; 3918c6fd2807SJeff Garzik 3919936fd732STejun Heo /* if SATA, resume link */ 3920a16abc0bSTejun Heo if (ap->flags & ATA_FLAG_SATA) { 3921936fd732STejun Heo rc = sata_link_resume(link, timing, deadline); 3922b8cffc6aSTejun Heo /* whine about phy resume failure but proceed */ 3923b8cffc6aSTejun Heo if (rc && rc != -EOPNOTSUPP) 3924cc0680a5STejun Heo ata_link_printk(link, KERN_WARNING, "failed to resume " 3925c6fd2807SJeff Garzik "link for reset (errno=%d)\n", rc); 3926c6fd2807SJeff Garzik } 3927c6fd2807SJeff Garzik 3928c6fd2807SJeff Garzik /* Wait for !BSY if the controller can wait for the first D2H 3929c6fd2807SJeff Garzik * Reg FIS and we don't know that no device is attached. 3930c6fd2807SJeff Garzik */ 39310c88758bSTejun Heo if (!(link->flags & ATA_LFLAG_SKIP_D2H_BSY) && !ata_link_offline(link)) { 3932b8cffc6aSTejun Heo rc = ata_wait_ready(ap, deadline); 39336dffaf61STejun Heo if (rc && rc != -ENODEV) { 3934cc0680a5STejun Heo ata_link_printk(link, KERN_WARNING, "device not ready " 3935b8cffc6aSTejun Heo "(errno=%d), forcing hardreset\n", rc); 3936b8cffc6aSTejun Heo ehc->i.action |= ATA_EH_HARDRESET; 3937b8cffc6aSTejun Heo } 3938b8cffc6aSTejun Heo } 3939c6fd2807SJeff Garzik 3940c6fd2807SJeff Garzik return 0; 3941c6fd2807SJeff Garzik } 3942c6fd2807SJeff Garzik 3943c6fd2807SJeff Garzik /** 3944c6fd2807SJeff Garzik * ata_std_softreset - reset host port via ATA SRST 3945cc0680a5STejun Heo * @link: ATA link to reset 3946c6fd2807SJeff Garzik * @classes: resulting classes of attached devices 3947d4b2bab4STejun Heo * @deadline: deadline jiffies for the operation 3948c6fd2807SJeff Garzik * 3949c6fd2807SJeff Garzik * Reset host port using ATA SRST. 3950c6fd2807SJeff Garzik * 3951c6fd2807SJeff Garzik * LOCKING: 3952c6fd2807SJeff Garzik * Kernel thread context (may sleep) 3953c6fd2807SJeff Garzik * 3954c6fd2807SJeff Garzik * RETURNS: 3955c6fd2807SJeff Garzik * 0 on success, -errno otherwise. 3956c6fd2807SJeff Garzik */ 3957cc0680a5STejun Heo int ata_std_softreset(struct ata_link *link, unsigned int *classes, 3958d4b2bab4STejun Heo unsigned long deadline) 3959c6fd2807SJeff Garzik { 3960cc0680a5STejun Heo struct ata_port *ap = link->ap; 3961c6fd2807SJeff Garzik unsigned int slave_possible = ap->flags & ATA_FLAG_SLAVE_POSS; 3962d4b2bab4STejun Heo unsigned int devmask = 0; 3963d4b2bab4STejun Heo int rc; 3964c6fd2807SJeff Garzik u8 err; 3965c6fd2807SJeff Garzik 3966c6fd2807SJeff Garzik DPRINTK("ENTER\n"); 3967c6fd2807SJeff Garzik 3968936fd732STejun Heo if (ata_link_offline(link)) { 3969c6fd2807SJeff Garzik classes[0] = ATA_DEV_NONE; 3970c6fd2807SJeff Garzik goto out; 3971c6fd2807SJeff Garzik } 3972c6fd2807SJeff Garzik 3973c6fd2807SJeff Garzik /* determine if device 0/1 are present */ 3974c6fd2807SJeff Garzik if (ata_devchk(ap, 0)) 3975c6fd2807SJeff Garzik devmask |= (1 << 0); 3976c6fd2807SJeff Garzik if (slave_possible && ata_devchk(ap, 1)) 3977c6fd2807SJeff Garzik devmask |= (1 << 1); 3978c6fd2807SJeff Garzik 3979c6fd2807SJeff Garzik /* select device 0 again */ 3980c6fd2807SJeff Garzik ap->ops->dev_select(ap, 0); 3981c6fd2807SJeff Garzik 3982c6fd2807SJeff Garzik /* issue bus reset */ 3983c6fd2807SJeff Garzik DPRINTK("about to softreset, devmask=%x\n", devmask); 3984d4b2bab4STejun Heo rc = ata_bus_softreset(ap, devmask, deadline); 39859b89391cSTejun Heo /* if link is occupied, -ENODEV too is an error */ 3986936fd732STejun Heo if (rc && (rc != -ENODEV || sata_scr_valid(link))) { 3987cc0680a5STejun Heo ata_link_printk(link, KERN_ERR, "SRST failed (errno=%d)\n", rc); 3988d4b2bab4STejun Heo return rc; 3989c6fd2807SJeff Garzik } 3990c6fd2807SJeff Garzik 3991c6fd2807SJeff Garzik /* determine by signature whether we have ATA or ATAPI devices */ 39923f19859eSTejun Heo classes[0] = ata_dev_try_classify(&link->device[0], 39933f19859eSTejun Heo devmask & (1 << 0), &err); 3994c6fd2807SJeff Garzik if (slave_possible && err != 0x81) 39953f19859eSTejun Heo classes[1] = ata_dev_try_classify(&link->device[1], 39963f19859eSTejun Heo devmask & (1 << 1), &err); 3997c6fd2807SJeff Garzik 3998c6fd2807SJeff Garzik out: 3999c6fd2807SJeff Garzik DPRINTK("EXIT, classes[0]=%u [1]=%u\n", classes[0], classes[1]); 4000c6fd2807SJeff Garzik return 0; 4001c6fd2807SJeff Garzik } 4002c6fd2807SJeff Garzik 4003c6fd2807SJeff Garzik /** 4004cc0680a5STejun Heo * sata_link_hardreset - reset link via SATA phy reset 4005cc0680a5STejun Heo * @link: link to reset 4006b6103f6dSTejun Heo * @timing: timing parameters { interval, duratinon, timeout } in msec 4007d4b2bab4STejun Heo * @deadline: deadline jiffies for the operation 4008c6fd2807SJeff Garzik * 4009cc0680a5STejun Heo * SATA phy-reset @link using DET bits of SControl register. 4010c6fd2807SJeff Garzik * 4011c6fd2807SJeff Garzik * LOCKING: 4012c6fd2807SJeff Garzik * Kernel thread context (may sleep) 4013c6fd2807SJeff Garzik * 4014c6fd2807SJeff Garzik * RETURNS: 4015c6fd2807SJeff Garzik * 0 on success, -errno otherwise. 4016c6fd2807SJeff Garzik */ 4017cc0680a5STejun Heo int sata_link_hardreset(struct ata_link *link, const unsigned long *timing, 4018d4b2bab4STejun Heo unsigned long deadline) 4019c6fd2807SJeff Garzik { 4020c6fd2807SJeff Garzik u32 scontrol; 4021c6fd2807SJeff Garzik int rc; 4022c6fd2807SJeff Garzik 4023c6fd2807SJeff Garzik DPRINTK("ENTER\n"); 4024c6fd2807SJeff Garzik 4025936fd732STejun Heo if (sata_set_spd_needed(link)) { 4026c6fd2807SJeff Garzik /* SATA spec says nothing about how to reconfigure 4027c6fd2807SJeff Garzik * spd. To be on the safe side, turn off phy during 4028c6fd2807SJeff Garzik * reconfiguration. This works for at least ICH7 AHCI 4029c6fd2807SJeff Garzik * and Sil3124. 4030c6fd2807SJeff Garzik */ 4031936fd732STejun Heo if ((rc = sata_scr_read(link, SCR_CONTROL, &scontrol))) 4032b6103f6dSTejun Heo goto out; 4033c6fd2807SJeff Garzik 4034cea0d336SJeff Garzik scontrol = (scontrol & 0x0f0) | 0x304; 4035c6fd2807SJeff Garzik 4036936fd732STejun Heo if ((rc = sata_scr_write(link, SCR_CONTROL, scontrol))) 4037b6103f6dSTejun Heo goto out; 4038c6fd2807SJeff Garzik 4039936fd732STejun Heo sata_set_spd(link); 4040c6fd2807SJeff Garzik } 4041c6fd2807SJeff Garzik 4042c6fd2807SJeff Garzik /* issue phy wake/reset */ 4043936fd732STejun Heo if ((rc = sata_scr_read(link, SCR_CONTROL, &scontrol))) 4044b6103f6dSTejun Heo goto out; 4045c6fd2807SJeff Garzik 4046c6fd2807SJeff Garzik scontrol = (scontrol & 0x0f0) | 0x301; 4047c6fd2807SJeff Garzik 4048936fd732STejun Heo if ((rc = sata_scr_write_flush(link, SCR_CONTROL, scontrol))) 4049b6103f6dSTejun Heo goto out; 4050c6fd2807SJeff Garzik 4051c6fd2807SJeff Garzik /* Couldn't find anything in SATA I/II specs, but AHCI-1.1 4052c6fd2807SJeff Garzik * 10.4.2 says at least 1 ms. 4053c6fd2807SJeff Garzik */ 4054c6fd2807SJeff Garzik msleep(1); 4055c6fd2807SJeff Garzik 4056936fd732STejun Heo /* bring link back */ 4057936fd732STejun Heo rc = sata_link_resume(link, timing, deadline); 4058b6103f6dSTejun Heo out: 4059b6103f6dSTejun Heo DPRINTK("EXIT, rc=%d\n", rc); 4060b6103f6dSTejun Heo return rc; 4061b6103f6dSTejun Heo } 4062b6103f6dSTejun Heo 4063b6103f6dSTejun Heo /** 4064b6103f6dSTejun Heo * sata_std_hardreset - reset host port via SATA phy reset 4065cc0680a5STejun Heo * @link: link to reset 4066b6103f6dSTejun Heo * @class: resulting class of attached device 4067d4b2bab4STejun Heo * @deadline: deadline jiffies for the operation 4068b6103f6dSTejun Heo * 4069b6103f6dSTejun Heo * SATA phy-reset host port using DET bits of SControl register, 4070b6103f6dSTejun Heo * wait for !BSY and classify the attached device. 4071b6103f6dSTejun Heo * 4072b6103f6dSTejun Heo * LOCKING: 4073b6103f6dSTejun Heo * Kernel thread context (may sleep) 4074b6103f6dSTejun Heo * 4075b6103f6dSTejun Heo * RETURNS: 4076b6103f6dSTejun Heo * 0 on success, -errno otherwise. 4077b6103f6dSTejun Heo */ 4078cc0680a5STejun Heo int sata_std_hardreset(struct ata_link *link, unsigned int *class, 4079d4b2bab4STejun Heo unsigned long deadline) 4080b6103f6dSTejun Heo { 4081cc0680a5STejun Heo struct ata_port *ap = link->ap; 4082936fd732STejun Heo const unsigned long *timing = sata_ehc_deb_timing(&link->eh_context); 4083b6103f6dSTejun Heo int rc; 4084b6103f6dSTejun Heo 4085b6103f6dSTejun Heo DPRINTK("ENTER\n"); 4086b6103f6dSTejun Heo 4087b6103f6dSTejun Heo /* do hardreset */ 4088cc0680a5STejun Heo rc = sata_link_hardreset(link, timing, deadline); 4089b6103f6dSTejun Heo if (rc) { 4090cc0680a5STejun Heo ata_link_printk(link, KERN_ERR, 4091b6103f6dSTejun Heo "COMRESET failed (errno=%d)\n", rc); 4092b6103f6dSTejun Heo return rc; 4093b6103f6dSTejun Heo } 4094c6fd2807SJeff Garzik 4095c6fd2807SJeff Garzik /* TODO: phy layer with polling, timeouts, etc. */ 4096936fd732STejun Heo if (ata_link_offline(link)) { 4097c6fd2807SJeff Garzik *class = ATA_DEV_NONE; 4098c6fd2807SJeff Garzik DPRINTK("EXIT, link offline\n"); 4099c6fd2807SJeff Garzik return 0; 4100c6fd2807SJeff Garzik } 4101c6fd2807SJeff Garzik 410288ff6eafSTejun Heo /* wait a while before checking status */ 410388ff6eafSTejun Heo ata_wait_after_reset(ap, deadline); 410434fee227STejun Heo 4105633273a3STejun Heo /* If PMP is supported, we have to do follow-up SRST. Note 4106633273a3STejun Heo * that some PMPs don't send D2H Reg FIS after hardreset at 4107633273a3STejun Heo * all if the first port is empty. Wait for it just for a 4108633273a3STejun Heo * second and request follow-up SRST. 4109633273a3STejun Heo */ 4110633273a3STejun Heo if (ap->flags & ATA_FLAG_PMP) { 4111633273a3STejun Heo ata_wait_ready(ap, jiffies + HZ); 4112633273a3STejun Heo return -EAGAIN; 4113633273a3STejun Heo } 4114633273a3STejun Heo 4115d4b2bab4STejun Heo rc = ata_wait_ready(ap, deadline); 41169b89391cSTejun Heo /* link occupied, -ENODEV too is an error */ 41179b89391cSTejun Heo if (rc) { 4118cc0680a5STejun Heo ata_link_printk(link, KERN_ERR, 4119d4b2bab4STejun Heo "COMRESET failed (errno=%d)\n", rc); 4120d4b2bab4STejun Heo return rc; 4121c6fd2807SJeff Garzik } 4122c6fd2807SJeff Garzik 4123c6fd2807SJeff Garzik ap->ops->dev_select(ap, 0); /* probably unnecessary */ 4124c6fd2807SJeff Garzik 41253f19859eSTejun Heo *class = ata_dev_try_classify(link->device, 1, NULL); 4126c6fd2807SJeff Garzik 4127c6fd2807SJeff Garzik DPRINTK("EXIT, class=%u\n", *class); 4128c6fd2807SJeff Garzik return 0; 4129c6fd2807SJeff Garzik } 4130c6fd2807SJeff Garzik 4131c6fd2807SJeff Garzik /** 4132c6fd2807SJeff Garzik * ata_std_postreset - standard postreset callback 4133cc0680a5STejun Heo * @link: the target ata_link 4134c6fd2807SJeff Garzik * @classes: classes of attached devices 4135c6fd2807SJeff Garzik * 4136c6fd2807SJeff Garzik * This function is invoked after a successful reset. Note that 4137c6fd2807SJeff Garzik * the device might have been reset more than once using 4138c6fd2807SJeff Garzik * different reset methods before postreset is invoked. 4139c6fd2807SJeff Garzik * 4140c6fd2807SJeff Garzik * LOCKING: 4141c6fd2807SJeff Garzik * Kernel thread context (may sleep) 4142c6fd2807SJeff Garzik */ 4143cc0680a5STejun Heo void ata_std_postreset(struct ata_link *link, unsigned int *classes) 4144c6fd2807SJeff Garzik { 4145cc0680a5STejun Heo struct ata_port *ap = link->ap; 4146c6fd2807SJeff Garzik u32 serror; 4147c6fd2807SJeff Garzik 4148c6fd2807SJeff Garzik DPRINTK("ENTER\n"); 4149c6fd2807SJeff Garzik 4150c6fd2807SJeff Garzik /* print link status */ 4151936fd732STejun Heo sata_print_link_status(link); 4152c6fd2807SJeff Garzik 4153c6fd2807SJeff Garzik /* clear SError */ 4154936fd732STejun Heo if (sata_scr_read(link, SCR_ERROR, &serror) == 0) 4155936fd732STejun Heo sata_scr_write(link, SCR_ERROR, serror); 4156f7fe7ad4STejun Heo link->eh_info.serror = 0; 4157c6fd2807SJeff Garzik 4158c6fd2807SJeff Garzik /* is double-select really necessary? */ 4159c6fd2807SJeff Garzik if (classes[0] != ATA_DEV_NONE) 4160c6fd2807SJeff Garzik ap->ops->dev_select(ap, 1); 4161c6fd2807SJeff Garzik if (classes[1] != ATA_DEV_NONE) 4162c6fd2807SJeff Garzik ap->ops->dev_select(ap, 0); 4163c6fd2807SJeff Garzik 4164c6fd2807SJeff Garzik /* bail out if no device is present */ 4165c6fd2807SJeff Garzik if (classes[0] == ATA_DEV_NONE && classes[1] == ATA_DEV_NONE) { 4166c6fd2807SJeff Garzik DPRINTK("EXIT, no device\n"); 4167c6fd2807SJeff Garzik return; 4168c6fd2807SJeff Garzik } 4169c6fd2807SJeff Garzik 4170c6fd2807SJeff Garzik /* set up device control */ 41710d5ff566STejun Heo if (ap->ioaddr.ctl_addr) 41720d5ff566STejun Heo iowrite8(ap->ctl, ap->ioaddr.ctl_addr); 4173c6fd2807SJeff Garzik 4174c6fd2807SJeff Garzik DPRINTK("EXIT\n"); 4175c6fd2807SJeff Garzik } 4176c6fd2807SJeff Garzik 4177c6fd2807SJeff Garzik /** 4178c6fd2807SJeff Garzik * ata_dev_same_device - Determine whether new ID matches configured device 4179c6fd2807SJeff Garzik * @dev: device to compare against 4180c6fd2807SJeff Garzik * @new_class: class of the new device 4181c6fd2807SJeff Garzik * @new_id: IDENTIFY page of the new device 4182c6fd2807SJeff Garzik * 4183c6fd2807SJeff Garzik * Compare @new_class and @new_id against @dev and determine 4184c6fd2807SJeff Garzik * whether @dev is the device indicated by @new_class and 4185c6fd2807SJeff Garzik * @new_id. 4186c6fd2807SJeff Garzik * 4187c6fd2807SJeff Garzik * LOCKING: 4188c6fd2807SJeff Garzik * None. 4189c6fd2807SJeff Garzik * 4190c6fd2807SJeff Garzik * RETURNS: 4191c6fd2807SJeff Garzik * 1 if @dev matches @new_class and @new_id, 0 otherwise. 4192c6fd2807SJeff Garzik */ 4193c6fd2807SJeff Garzik static int ata_dev_same_device(struct ata_device *dev, unsigned int new_class, 4194c6fd2807SJeff Garzik const u16 *new_id) 4195c6fd2807SJeff Garzik { 4196c6fd2807SJeff Garzik const u16 *old_id = dev->id; 4197a0cf733bSTejun Heo unsigned char model[2][ATA_ID_PROD_LEN + 1]; 4198a0cf733bSTejun Heo unsigned char serial[2][ATA_ID_SERNO_LEN + 1]; 4199c6fd2807SJeff Garzik 4200c6fd2807SJeff Garzik if (dev->class != new_class) { 4201c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_INFO, "class mismatch %d != %d\n", 4202c6fd2807SJeff Garzik dev->class, new_class); 4203c6fd2807SJeff Garzik return 0; 4204c6fd2807SJeff Garzik } 4205c6fd2807SJeff Garzik 4206a0cf733bSTejun Heo ata_id_c_string(old_id, model[0], ATA_ID_PROD, sizeof(model[0])); 4207a0cf733bSTejun Heo ata_id_c_string(new_id, model[1], ATA_ID_PROD, sizeof(model[1])); 4208a0cf733bSTejun Heo ata_id_c_string(old_id, serial[0], ATA_ID_SERNO, sizeof(serial[0])); 4209a0cf733bSTejun Heo ata_id_c_string(new_id, serial[1], ATA_ID_SERNO, sizeof(serial[1])); 4210c6fd2807SJeff Garzik 4211c6fd2807SJeff Garzik if (strcmp(model[0], model[1])) { 4212c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_INFO, "model number mismatch " 4213c6fd2807SJeff Garzik "'%s' != '%s'\n", model[0], model[1]); 4214c6fd2807SJeff Garzik return 0; 4215c6fd2807SJeff Garzik } 4216c6fd2807SJeff Garzik 4217c6fd2807SJeff Garzik if (strcmp(serial[0], serial[1])) { 4218c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_INFO, "serial number mismatch " 4219c6fd2807SJeff Garzik "'%s' != '%s'\n", serial[0], serial[1]); 4220c6fd2807SJeff Garzik return 0; 4221c6fd2807SJeff Garzik } 4222c6fd2807SJeff Garzik 4223c6fd2807SJeff Garzik return 1; 4224c6fd2807SJeff Garzik } 4225c6fd2807SJeff Garzik 4226c6fd2807SJeff Garzik /** 4227fe30911bSTejun Heo * ata_dev_reread_id - Re-read IDENTIFY data 42283fae450cSHenrik Kretzschmar * @dev: target ATA device 4229bff04647STejun Heo * @readid_flags: read ID flags 4230c6fd2807SJeff Garzik * 4231c6fd2807SJeff Garzik * Re-read IDENTIFY page and make sure @dev is still attached to 4232c6fd2807SJeff Garzik * the port. 4233c6fd2807SJeff Garzik * 4234c6fd2807SJeff Garzik * LOCKING: 4235c6fd2807SJeff Garzik * Kernel thread context (may sleep) 4236c6fd2807SJeff Garzik * 4237c6fd2807SJeff Garzik * RETURNS: 4238c6fd2807SJeff Garzik * 0 on success, negative errno otherwise 4239c6fd2807SJeff Garzik */ 4240fe30911bSTejun Heo int ata_dev_reread_id(struct ata_device *dev, unsigned int readid_flags) 4241c6fd2807SJeff Garzik { 4242c6fd2807SJeff Garzik unsigned int class = dev->class; 42439af5c9c9STejun Heo u16 *id = (void *)dev->link->ap->sector_buf; 4244c6fd2807SJeff Garzik int rc; 4245c6fd2807SJeff Garzik 4246c6fd2807SJeff Garzik /* read ID data */ 4247bff04647STejun Heo rc = ata_dev_read_id(dev, &class, readid_flags, id); 4248c6fd2807SJeff Garzik if (rc) 4249fe30911bSTejun Heo return rc; 4250c6fd2807SJeff Garzik 4251c6fd2807SJeff Garzik /* is the device still there? */ 4252fe30911bSTejun Heo if (!ata_dev_same_device(dev, class, id)) 4253fe30911bSTejun Heo return -ENODEV; 4254c6fd2807SJeff Garzik 4255c6fd2807SJeff Garzik memcpy(dev->id, id, sizeof(id[0]) * ATA_ID_WORDS); 4256fe30911bSTejun Heo return 0; 4257fe30911bSTejun Heo } 4258fe30911bSTejun Heo 4259fe30911bSTejun Heo /** 4260fe30911bSTejun Heo * ata_dev_revalidate - Revalidate ATA device 4261fe30911bSTejun Heo * @dev: device to revalidate 4262422c9daaSTejun Heo * @new_class: new class code 4263fe30911bSTejun Heo * @readid_flags: read ID flags 4264fe30911bSTejun Heo * 4265fe30911bSTejun Heo * Re-read IDENTIFY page, make sure @dev is still attached to the 4266fe30911bSTejun Heo * port and reconfigure it according to the new IDENTIFY page. 4267fe30911bSTejun Heo * 4268fe30911bSTejun Heo * LOCKING: 4269fe30911bSTejun Heo * Kernel thread context (may sleep) 4270fe30911bSTejun Heo * 4271fe30911bSTejun Heo * RETURNS: 4272fe30911bSTejun Heo * 0 on success, negative errno otherwise 4273fe30911bSTejun Heo */ 4274422c9daaSTejun Heo int ata_dev_revalidate(struct ata_device *dev, unsigned int new_class, 4275422c9daaSTejun Heo unsigned int readid_flags) 4276fe30911bSTejun Heo { 42776ddcd3b0STejun Heo u64 n_sectors = dev->n_sectors; 4278fe30911bSTejun Heo int rc; 4279fe30911bSTejun Heo 4280fe30911bSTejun Heo if (!ata_dev_enabled(dev)) 4281fe30911bSTejun Heo return -ENODEV; 4282fe30911bSTejun Heo 4283422c9daaSTejun Heo /* fail early if !ATA && !ATAPI to avoid issuing [P]IDENTIFY to PMP */ 4284422c9daaSTejun Heo if (ata_class_enabled(new_class) && 4285422c9daaSTejun Heo new_class != ATA_DEV_ATA && new_class != ATA_DEV_ATAPI) { 4286422c9daaSTejun Heo ata_dev_printk(dev, KERN_INFO, "class mismatch %u != %u\n", 4287422c9daaSTejun Heo dev->class, new_class); 4288422c9daaSTejun Heo rc = -ENODEV; 4289422c9daaSTejun Heo goto fail; 4290422c9daaSTejun Heo } 4291422c9daaSTejun Heo 4292fe30911bSTejun Heo /* re-read ID */ 4293fe30911bSTejun Heo rc = ata_dev_reread_id(dev, readid_flags); 4294fe30911bSTejun Heo if (rc) 4295fe30911bSTejun Heo goto fail; 4296c6fd2807SJeff Garzik 4297c6fd2807SJeff Garzik /* configure device according to the new ID */ 4298efdaedc4STejun Heo rc = ata_dev_configure(dev); 42996ddcd3b0STejun Heo if (rc) 43006ddcd3b0STejun Heo goto fail; 43016ddcd3b0STejun Heo 43026ddcd3b0STejun Heo /* verify n_sectors hasn't changed */ 4303b54eebd6STejun Heo if (dev->class == ATA_DEV_ATA && n_sectors && 4304b54eebd6STejun Heo dev->n_sectors != n_sectors) { 43056ddcd3b0STejun Heo ata_dev_printk(dev, KERN_INFO, "n_sectors mismatch " 43066ddcd3b0STejun Heo "%llu != %llu\n", 43076ddcd3b0STejun Heo (unsigned long long)n_sectors, 43086ddcd3b0STejun Heo (unsigned long long)dev->n_sectors); 43098270bec4STejun Heo 43108270bec4STejun Heo /* restore original n_sectors */ 43118270bec4STejun Heo dev->n_sectors = n_sectors; 43128270bec4STejun Heo 43136ddcd3b0STejun Heo rc = -ENODEV; 43146ddcd3b0STejun Heo goto fail; 43156ddcd3b0STejun Heo } 43166ddcd3b0STejun Heo 4317c6fd2807SJeff Garzik return 0; 4318c6fd2807SJeff Garzik 4319c6fd2807SJeff Garzik fail: 4320c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_ERR, "revalidation failed (errno=%d)\n", rc); 4321c6fd2807SJeff Garzik return rc; 4322c6fd2807SJeff Garzik } 4323c6fd2807SJeff Garzik 43246919a0a6SAlan Cox struct ata_blacklist_entry { 43256919a0a6SAlan Cox const char *model_num; 43266919a0a6SAlan Cox const char *model_rev; 43276919a0a6SAlan Cox unsigned long horkage; 43286919a0a6SAlan Cox }; 43296919a0a6SAlan Cox 43306919a0a6SAlan Cox static const struct ata_blacklist_entry ata_device_blacklist [] = { 43316919a0a6SAlan Cox /* Devices with DMA related problems under Linux */ 43326919a0a6SAlan Cox { "WDC AC11000H", NULL, ATA_HORKAGE_NODMA }, 43336919a0a6SAlan Cox { "WDC AC22100H", NULL, ATA_HORKAGE_NODMA }, 43346919a0a6SAlan Cox { "WDC AC32500H", NULL, ATA_HORKAGE_NODMA }, 43356919a0a6SAlan Cox { "WDC AC33100H", NULL, ATA_HORKAGE_NODMA }, 43366919a0a6SAlan Cox { "WDC AC31600H", NULL, ATA_HORKAGE_NODMA }, 43376919a0a6SAlan Cox { "WDC AC32100H", "24.09P07", ATA_HORKAGE_NODMA }, 43386919a0a6SAlan Cox { "WDC AC23200L", "21.10N21", ATA_HORKAGE_NODMA }, 43396919a0a6SAlan Cox { "Compaq CRD-8241B", NULL, ATA_HORKAGE_NODMA }, 43406919a0a6SAlan Cox { "CRD-8400B", NULL, ATA_HORKAGE_NODMA }, 43416919a0a6SAlan Cox { "CRD-8480B", NULL, ATA_HORKAGE_NODMA }, 43426919a0a6SAlan Cox { "CRD-8482B", NULL, ATA_HORKAGE_NODMA }, 43436919a0a6SAlan Cox { "CRD-84", NULL, ATA_HORKAGE_NODMA }, 43446919a0a6SAlan Cox { "SanDisk SDP3B", NULL, ATA_HORKAGE_NODMA }, 43456919a0a6SAlan Cox { "SanDisk SDP3B-64", NULL, ATA_HORKAGE_NODMA }, 43466919a0a6SAlan Cox { "SANYO CD-ROM CRD", NULL, ATA_HORKAGE_NODMA }, 43476919a0a6SAlan Cox { "HITACHI CDR-8", NULL, ATA_HORKAGE_NODMA }, 43486919a0a6SAlan Cox { "HITACHI CDR-8335", NULL, ATA_HORKAGE_NODMA }, 43496919a0a6SAlan Cox { "HITACHI CDR-8435", NULL, ATA_HORKAGE_NODMA }, 43506919a0a6SAlan Cox { "Toshiba CD-ROM XM-6202B", NULL, ATA_HORKAGE_NODMA }, 43516919a0a6SAlan Cox { "TOSHIBA CD-ROM XM-1702BC", NULL, ATA_HORKAGE_NODMA }, 43526919a0a6SAlan Cox { "CD-532E-A", NULL, ATA_HORKAGE_NODMA }, 43536919a0a6SAlan Cox { "E-IDE CD-ROM CR-840",NULL, ATA_HORKAGE_NODMA }, 43546919a0a6SAlan Cox { "CD-ROM Drive/F5A", NULL, ATA_HORKAGE_NODMA }, 43556919a0a6SAlan Cox { "WPI CDD-820", NULL, ATA_HORKAGE_NODMA }, 43566919a0a6SAlan Cox { "SAMSUNG CD-ROM SC-148C", NULL, ATA_HORKAGE_NODMA }, 43576919a0a6SAlan Cox { "SAMSUNG CD-ROM SC", NULL, ATA_HORKAGE_NODMA }, 43586919a0a6SAlan Cox { "ATAPI CD-ROM DRIVE 40X MAXIMUM",NULL,ATA_HORKAGE_NODMA }, 43596919a0a6SAlan Cox { "_NEC DV5800A", NULL, ATA_HORKAGE_NODMA }, 43606919a0a6SAlan Cox { "SAMSUNG CD-ROM SN-124", "N001", ATA_HORKAGE_NODMA }, 436139f19886SDave Jones { "Seagate STT20000A", NULL, ATA_HORKAGE_NODMA }, 43623af9a77aSTejun Heo /* Odd clown on sil3726/4726 PMPs */ 43633af9a77aSTejun Heo { "Config Disk", NULL, ATA_HORKAGE_NODMA | 43643af9a77aSTejun Heo ATA_HORKAGE_SKIP_PM }, 43656919a0a6SAlan Cox 436618d6e9d5SAlbert Lee /* Weird ATAPI devices */ 436740a1d531STejun Heo { "TORiSAN DVD-ROM DRD-N216", NULL, ATA_HORKAGE_MAX_SEC_128 }, 436818d6e9d5SAlbert Lee 43696919a0a6SAlan Cox /* Devices we expect to fail diagnostics */ 43706919a0a6SAlan Cox 43716919a0a6SAlan Cox /* Devices where NCQ should be avoided */ 43726919a0a6SAlan Cox /* NCQ is slow */ 43736919a0a6SAlan Cox { "WDC WD740ADFD-00", NULL, ATA_HORKAGE_NONCQ }, 4374459ad688STejun Heo { "WDC WD740ADFD-00NLR1", NULL, ATA_HORKAGE_NONCQ, }, 437509125ea6STejun Heo /* http://thread.gmane.org/gmane.linux.ide/14907 */ 437609125ea6STejun Heo { "FUJITSU MHT2060BH", NULL, ATA_HORKAGE_NONCQ }, 43777acfaf30SPaul Rolland /* NCQ is broken */ 4378539cc7c7SJeff Garzik { "Maxtor *", "BANC*", ATA_HORKAGE_NONCQ }, 43790e3dbc01SAlan Cox { "Maxtor 7V300F0", "VA111630", ATA_HORKAGE_NONCQ }, 4380da6f0ec2SPaolo Ornati { "ST380817AS", "3.42", ATA_HORKAGE_NONCQ }, 4381e41bd3e8STejun Heo { "ST3160023AS", "3.42", ATA_HORKAGE_NONCQ }, 4382539cc7c7SJeff Garzik 438336e337d0SRobert Hancock /* Blacklist entries taken from Silicon Image 3124/3132 438436e337d0SRobert Hancock Windows driver .inf file - also several Linux problem reports */ 438536e337d0SRobert Hancock { "HTS541060G9SA00", "MB3OC60D", ATA_HORKAGE_NONCQ, }, 438636e337d0SRobert Hancock { "HTS541080G9SA00", "MB4OC60D", ATA_HORKAGE_NONCQ, }, 438736e337d0SRobert Hancock { "HTS541010G9SA00", "MBZOC60D", ATA_HORKAGE_NONCQ, }, 43886919a0a6SAlan Cox 438916c55b03STejun Heo /* devices which puke on READ_NATIVE_MAX */ 439016c55b03STejun Heo { "HDS724040KLSA80", "KFAOA20N", ATA_HORKAGE_BROKEN_HPA, }, 439116c55b03STejun Heo { "WDC WD3200JD-00KLB0", "WD-WCAMR1130137", ATA_HORKAGE_BROKEN_HPA }, 439216c55b03STejun Heo { "WDC WD2500JD-00HBB0", "WD-WMAL71490727", ATA_HORKAGE_BROKEN_HPA }, 439316c55b03STejun Heo { "MAXTOR 6L080L4", "A93.0500", ATA_HORKAGE_BROKEN_HPA }, 43946919a0a6SAlan Cox 439593328e11SAlan Cox /* Devices which report 1 sector over size HPA */ 439693328e11SAlan Cox { "ST340823A", NULL, ATA_HORKAGE_HPA_SIZE, }, 439793328e11SAlan Cox { "ST320413A", NULL, ATA_HORKAGE_HPA_SIZE, }, 4398b152fcd3SMikko Rapeli { "ST310211A", NULL, ATA_HORKAGE_HPA_SIZE, }, 439993328e11SAlan Cox 44006bbfd53dSAlan Cox /* Devices which get the IVB wrong */ 44016bbfd53dSAlan Cox { "QUANTUM FIREBALLlct10 05", "A03.0900", ATA_HORKAGE_IVB, }, 44026bbfd53dSAlan Cox { "TSSTcorp CDDVDW SH-S202J", "SB00", ATA_HORKAGE_IVB, }, 4403e9f33406SPeter Missel { "TSSTcorp CDDVDW SH-S202J", "SB01", ATA_HORKAGE_IVB, }, 4404e9f33406SPeter Missel { "TSSTcorp CDDVDW SH-S202N", "SB00", ATA_HORKAGE_IVB, }, 4405e9f33406SPeter Missel { "TSSTcorp CDDVDW SH-S202N", "SB01", ATA_HORKAGE_IVB, }, 44066bbfd53dSAlan Cox 44076919a0a6SAlan Cox /* End Marker */ 44086919a0a6SAlan Cox { } 4409c6fd2807SJeff Garzik }; 4410c6fd2807SJeff Garzik 4411741b7763SAdrian Bunk static int strn_pattern_cmp(const char *patt, const char *name, int wildchar) 4412539cc7c7SJeff Garzik { 4413539cc7c7SJeff Garzik const char *p; 4414539cc7c7SJeff Garzik int len; 4415539cc7c7SJeff Garzik 4416539cc7c7SJeff Garzik /* 4417539cc7c7SJeff Garzik * check for trailing wildcard: *\0 4418539cc7c7SJeff Garzik */ 4419539cc7c7SJeff Garzik p = strchr(patt, wildchar); 4420539cc7c7SJeff Garzik if (p && ((*(p + 1)) == 0)) 4421539cc7c7SJeff Garzik len = p - patt; 4422317b50b8SAndrew Paprocki else { 4423539cc7c7SJeff Garzik len = strlen(name); 4424317b50b8SAndrew Paprocki if (!len) { 4425317b50b8SAndrew Paprocki if (!*patt) 4426317b50b8SAndrew Paprocki return 0; 4427317b50b8SAndrew Paprocki return -1; 4428317b50b8SAndrew Paprocki } 4429317b50b8SAndrew Paprocki } 4430539cc7c7SJeff Garzik 4431539cc7c7SJeff Garzik return strncmp(patt, name, len); 4432539cc7c7SJeff Garzik } 4433539cc7c7SJeff Garzik 443475683fe7STejun Heo static unsigned long ata_dev_blacklisted(const struct ata_device *dev) 4435c6fd2807SJeff Garzik { 44368bfa79fcSTejun Heo unsigned char model_num[ATA_ID_PROD_LEN + 1]; 44378bfa79fcSTejun Heo unsigned char model_rev[ATA_ID_FW_REV_LEN + 1]; 44386919a0a6SAlan Cox const struct ata_blacklist_entry *ad = ata_device_blacklist; 4439c6fd2807SJeff Garzik 44408bfa79fcSTejun Heo ata_id_c_string(dev->id, model_num, ATA_ID_PROD, sizeof(model_num)); 44418bfa79fcSTejun Heo ata_id_c_string(dev->id, model_rev, ATA_ID_FW_REV, sizeof(model_rev)); 4442c6fd2807SJeff Garzik 44436919a0a6SAlan Cox while (ad->model_num) { 4444539cc7c7SJeff Garzik if (!strn_pattern_cmp(ad->model_num, model_num, '*')) { 44456919a0a6SAlan Cox if (ad->model_rev == NULL) 44466919a0a6SAlan Cox return ad->horkage; 4447539cc7c7SJeff Garzik if (!strn_pattern_cmp(ad->model_rev, model_rev, '*')) 44486919a0a6SAlan Cox return ad->horkage; 4449c6fd2807SJeff Garzik } 44506919a0a6SAlan Cox ad++; 4451c6fd2807SJeff Garzik } 4452c6fd2807SJeff Garzik return 0; 4453c6fd2807SJeff Garzik } 4454c6fd2807SJeff Garzik 44556919a0a6SAlan Cox static int ata_dma_blacklisted(const struct ata_device *dev) 44566919a0a6SAlan Cox { 44576919a0a6SAlan Cox /* We don't support polling DMA. 44586919a0a6SAlan Cox * DMA blacklist those ATAPI devices with CDB-intr (and use PIO) 44596919a0a6SAlan Cox * if the LLDD handles only interrupts in the HSM_ST_LAST state. 44606919a0a6SAlan Cox */ 44619af5c9c9STejun Heo if ((dev->link->ap->flags & ATA_FLAG_PIO_POLLING) && 44626919a0a6SAlan Cox (dev->flags & ATA_DFLAG_CDB_INTR)) 44636919a0a6SAlan Cox return 1; 446475683fe7STejun Heo return (dev->horkage & ATA_HORKAGE_NODMA) ? 1 : 0; 44656919a0a6SAlan Cox } 44666919a0a6SAlan Cox 4467c6fd2807SJeff Garzik /** 44686bbfd53dSAlan Cox * ata_is_40wire - check drive side detection 44696bbfd53dSAlan Cox * @dev: device 44706bbfd53dSAlan Cox * 44716bbfd53dSAlan Cox * Perform drive side detection decoding, allowing for device vendors 44726bbfd53dSAlan Cox * who can't follow the documentation. 44736bbfd53dSAlan Cox */ 44746bbfd53dSAlan Cox 44756bbfd53dSAlan Cox static int ata_is_40wire(struct ata_device *dev) 44766bbfd53dSAlan Cox { 44776bbfd53dSAlan Cox if (dev->horkage & ATA_HORKAGE_IVB) 44786bbfd53dSAlan Cox return ata_drive_40wire_relaxed(dev->id); 44796bbfd53dSAlan Cox return ata_drive_40wire(dev->id); 44806bbfd53dSAlan Cox } 44816bbfd53dSAlan Cox 44826bbfd53dSAlan Cox /** 4483c6fd2807SJeff Garzik * ata_dev_xfermask - Compute supported xfermask of the given device 4484c6fd2807SJeff Garzik * @dev: Device to compute xfermask for 4485c6fd2807SJeff Garzik * 4486c6fd2807SJeff Garzik * Compute supported xfermask of @dev and store it in 4487c6fd2807SJeff Garzik * dev->*_mask. This function is responsible for applying all 4488c6fd2807SJeff Garzik * known limits including host controller limits, device 4489c6fd2807SJeff Garzik * blacklist, etc... 4490c6fd2807SJeff Garzik * 4491c6fd2807SJeff Garzik * LOCKING: 4492c6fd2807SJeff Garzik * None. 4493c6fd2807SJeff Garzik */ 4494c6fd2807SJeff Garzik static void ata_dev_xfermask(struct ata_device *dev) 4495c6fd2807SJeff Garzik { 44969af5c9c9STejun Heo struct ata_link *link = dev->link; 44979af5c9c9STejun Heo struct ata_port *ap = link->ap; 4498cca3974eSJeff Garzik struct ata_host *host = ap->host; 4499c6fd2807SJeff Garzik unsigned long xfer_mask; 4500c6fd2807SJeff Garzik 4501c6fd2807SJeff Garzik /* controller modes available */ 4502c6fd2807SJeff Garzik xfer_mask = ata_pack_xfermask(ap->pio_mask, 4503c6fd2807SJeff Garzik ap->mwdma_mask, ap->udma_mask); 4504c6fd2807SJeff Garzik 45058343f889SRobert Hancock /* drive modes available */ 4506c6fd2807SJeff Garzik xfer_mask &= ata_pack_xfermask(dev->pio_mask, 4507c6fd2807SJeff Garzik dev->mwdma_mask, dev->udma_mask); 4508c6fd2807SJeff Garzik xfer_mask &= ata_id_xfermask(dev->id); 4509c6fd2807SJeff Garzik 4510b352e57dSAlan Cox /* 4511b352e57dSAlan Cox * CFA Advanced TrueIDE timings are not allowed on a shared 4512b352e57dSAlan Cox * cable 4513b352e57dSAlan Cox */ 4514b352e57dSAlan Cox if (ata_dev_pair(dev)) { 4515b352e57dSAlan Cox /* No PIO5 or PIO6 */ 4516b352e57dSAlan Cox xfer_mask &= ~(0x03 << (ATA_SHIFT_PIO + 5)); 4517b352e57dSAlan Cox /* No MWDMA3 or MWDMA 4 */ 4518b352e57dSAlan Cox xfer_mask &= ~(0x03 << (ATA_SHIFT_MWDMA + 3)); 4519b352e57dSAlan Cox } 4520b352e57dSAlan Cox 4521c6fd2807SJeff Garzik if (ata_dma_blacklisted(dev)) { 4522c6fd2807SJeff Garzik xfer_mask &= ~(ATA_MASK_MWDMA | ATA_MASK_UDMA); 4523c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_WARNING, 4524c6fd2807SJeff Garzik "device is on DMA blacklist, disabling DMA\n"); 4525c6fd2807SJeff Garzik } 4526c6fd2807SJeff Garzik 452714d66ab7SPetr Vandrovec if ((host->flags & ATA_HOST_SIMPLEX) && 452814d66ab7SPetr Vandrovec host->simplex_claimed && host->simplex_claimed != ap) { 4529c6fd2807SJeff Garzik xfer_mask &= ~(ATA_MASK_MWDMA | ATA_MASK_UDMA); 4530c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_WARNING, "simplex DMA is claimed by " 4531c6fd2807SJeff Garzik "other device, disabling DMA\n"); 4532c6fd2807SJeff Garzik } 4533c6fd2807SJeff Garzik 4534e424675fSJeff Garzik if (ap->flags & ATA_FLAG_NO_IORDY) 4535e424675fSJeff Garzik xfer_mask &= ata_pio_mask_no_iordy(dev); 4536e424675fSJeff Garzik 4537c6fd2807SJeff Garzik if (ap->ops->mode_filter) 4538a76b62caSAlan Cox xfer_mask = ap->ops->mode_filter(dev, xfer_mask); 4539c6fd2807SJeff Garzik 45408343f889SRobert Hancock /* Apply cable rule here. Don't apply it early because when 45418343f889SRobert Hancock * we handle hot plug the cable type can itself change. 45428343f889SRobert Hancock * Check this last so that we know if the transfer rate was 45438343f889SRobert Hancock * solely limited by the cable. 45448343f889SRobert Hancock * Unknown or 80 wire cables reported host side are checked 45458343f889SRobert Hancock * drive side as well. Cases where we know a 40wire cable 45468343f889SRobert Hancock * is used safely for 80 are not checked here. 45478343f889SRobert Hancock */ 45488343f889SRobert Hancock if (xfer_mask & (0xF8 << ATA_SHIFT_UDMA)) 45498343f889SRobert Hancock /* UDMA/44 or higher would be available */ 45508343f889SRobert Hancock if ((ap->cbl == ATA_CBL_PATA40) || 45516bbfd53dSAlan Cox (ata_is_40wire(dev) && 45528343f889SRobert Hancock (ap->cbl == ATA_CBL_PATA_UNK || 45538343f889SRobert Hancock ap->cbl == ATA_CBL_PATA80))) { 45548343f889SRobert Hancock ata_dev_printk(dev, KERN_WARNING, 45558343f889SRobert Hancock "limited to UDMA/33 due to 40-wire cable\n"); 45568343f889SRobert Hancock xfer_mask &= ~(0xF8 << ATA_SHIFT_UDMA); 45578343f889SRobert Hancock } 45588343f889SRobert Hancock 4559c6fd2807SJeff Garzik ata_unpack_xfermask(xfer_mask, &dev->pio_mask, 4560c6fd2807SJeff Garzik &dev->mwdma_mask, &dev->udma_mask); 4561c6fd2807SJeff Garzik } 4562c6fd2807SJeff Garzik 4563c6fd2807SJeff Garzik /** 4564c6fd2807SJeff Garzik * ata_dev_set_xfermode - Issue SET FEATURES - XFER MODE command 4565c6fd2807SJeff Garzik * @dev: Device to which command will be sent 4566c6fd2807SJeff Garzik * 4567c6fd2807SJeff Garzik * Issue SET FEATURES - XFER MODE command to device @dev 4568c6fd2807SJeff Garzik * on port @ap. 4569c6fd2807SJeff Garzik * 4570c6fd2807SJeff Garzik * LOCKING: 4571c6fd2807SJeff Garzik * PCI/etc. bus probe sem. 4572c6fd2807SJeff Garzik * 4573c6fd2807SJeff Garzik * RETURNS: 4574c6fd2807SJeff Garzik * 0 on success, AC_ERR_* mask otherwise. 4575c6fd2807SJeff Garzik */ 4576c6fd2807SJeff Garzik 4577c6fd2807SJeff Garzik static unsigned int ata_dev_set_xfermode(struct ata_device *dev) 4578c6fd2807SJeff Garzik { 4579c6fd2807SJeff Garzik struct ata_taskfile tf; 4580c6fd2807SJeff Garzik unsigned int err_mask; 4581c6fd2807SJeff Garzik 4582c6fd2807SJeff Garzik /* set up set-features taskfile */ 4583c6fd2807SJeff Garzik DPRINTK("set features - xfer mode\n"); 4584c6fd2807SJeff Garzik 4585464cf177STejun Heo /* Some controllers and ATAPI devices show flaky interrupt 4586464cf177STejun Heo * behavior after setting xfer mode. Use polling instead. 4587464cf177STejun Heo */ 4588c6fd2807SJeff Garzik ata_tf_init(dev, &tf); 4589c6fd2807SJeff Garzik tf.command = ATA_CMD_SET_FEATURES; 4590c6fd2807SJeff Garzik tf.feature = SETFEATURES_XFER; 4591464cf177STejun Heo tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE | ATA_TFLAG_POLLING; 4592c6fd2807SJeff Garzik tf.protocol = ATA_PROT_NODATA; 4593b9f8ab2dSAlan Cox /* If we are using IORDY we must send the mode setting command */ 4594b9f8ab2dSAlan Cox if (ata_pio_need_iordy(dev)) 4595c6fd2807SJeff Garzik tf.nsect = dev->xfer_mode; 4596b9f8ab2dSAlan Cox /* If the device has IORDY and the controller does not - turn it off */ 4597b9f8ab2dSAlan Cox else if (ata_id_has_iordy(dev->id)) 4598b9f8ab2dSAlan Cox tf.nsect = 0x01; 4599b9f8ab2dSAlan Cox else /* In the ancient relic department - skip all of this */ 4600b9f8ab2dSAlan Cox return 0; 4601c6fd2807SJeff Garzik 46022b789108STejun Heo err_mask = ata_exec_internal(dev, &tf, NULL, DMA_NONE, NULL, 0, 0); 4603c6fd2807SJeff Garzik 4604c6fd2807SJeff Garzik DPRINTK("EXIT, err_mask=%x\n", err_mask); 4605c6fd2807SJeff Garzik return err_mask; 4606c6fd2807SJeff Garzik } 4607c6fd2807SJeff Garzik /** 4608218f3d30SJeff Garzik * ata_dev_set_feature - Issue SET FEATURES - SATA FEATURES 46099f45cbd3SKristen Carlson Accardi * @dev: Device to which command will be sent 46109f45cbd3SKristen Carlson Accardi * @enable: Whether to enable or disable the feature 4611218f3d30SJeff Garzik * @feature: The sector count represents the feature to set 46129f45cbd3SKristen Carlson Accardi * 46139f45cbd3SKristen Carlson Accardi * Issue SET FEATURES - SATA FEATURES command to device @dev 4614218f3d30SJeff Garzik * on port @ap with sector count 46159f45cbd3SKristen Carlson Accardi * 46169f45cbd3SKristen Carlson Accardi * LOCKING: 46179f45cbd3SKristen Carlson Accardi * PCI/etc. bus probe sem. 46189f45cbd3SKristen Carlson Accardi * 46199f45cbd3SKristen Carlson Accardi * RETURNS: 46209f45cbd3SKristen Carlson Accardi * 0 on success, AC_ERR_* mask otherwise. 46219f45cbd3SKristen Carlson Accardi */ 4622218f3d30SJeff Garzik static unsigned int ata_dev_set_feature(struct ata_device *dev, u8 enable, 4623218f3d30SJeff Garzik u8 feature) 46249f45cbd3SKristen Carlson Accardi { 46259f45cbd3SKristen Carlson Accardi struct ata_taskfile tf; 46269f45cbd3SKristen Carlson Accardi unsigned int err_mask; 46279f45cbd3SKristen Carlson Accardi 46289f45cbd3SKristen Carlson Accardi /* set up set-features taskfile */ 46299f45cbd3SKristen Carlson Accardi DPRINTK("set features - SATA features\n"); 46309f45cbd3SKristen Carlson Accardi 46319f45cbd3SKristen Carlson Accardi ata_tf_init(dev, &tf); 46329f45cbd3SKristen Carlson Accardi tf.command = ATA_CMD_SET_FEATURES; 46339f45cbd3SKristen Carlson Accardi tf.feature = enable; 46349f45cbd3SKristen Carlson Accardi tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; 46359f45cbd3SKristen Carlson Accardi tf.protocol = ATA_PROT_NODATA; 4636218f3d30SJeff Garzik tf.nsect = feature; 46379f45cbd3SKristen Carlson Accardi 46382b789108STejun Heo err_mask = ata_exec_internal(dev, &tf, NULL, DMA_NONE, NULL, 0, 0); 46399f45cbd3SKristen Carlson Accardi 46409f45cbd3SKristen Carlson Accardi DPRINTK("EXIT, err_mask=%x\n", err_mask); 46419f45cbd3SKristen Carlson Accardi return err_mask; 46429f45cbd3SKristen Carlson Accardi } 46439f45cbd3SKristen Carlson Accardi 46449f45cbd3SKristen Carlson Accardi /** 4645c6fd2807SJeff Garzik * ata_dev_init_params - Issue INIT DEV PARAMS command 4646c6fd2807SJeff Garzik * @dev: Device to which command will be sent 4647c6fd2807SJeff Garzik * @heads: Number of heads (taskfile parameter) 4648c6fd2807SJeff Garzik * @sectors: Number of sectors (taskfile parameter) 4649c6fd2807SJeff Garzik * 4650c6fd2807SJeff Garzik * LOCKING: 4651c6fd2807SJeff Garzik * Kernel thread context (may sleep) 4652c6fd2807SJeff Garzik * 4653c6fd2807SJeff Garzik * RETURNS: 4654c6fd2807SJeff Garzik * 0 on success, AC_ERR_* mask otherwise. 4655c6fd2807SJeff Garzik */ 4656c6fd2807SJeff Garzik static unsigned int ata_dev_init_params(struct ata_device *dev, 4657c6fd2807SJeff Garzik u16 heads, u16 sectors) 4658c6fd2807SJeff Garzik { 4659c6fd2807SJeff Garzik struct ata_taskfile tf; 4660c6fd2807SJeff Garzik unsigned int err_mask; 4661c6fd2807SJeff Garzik 4662c6fd2807SJeff Garzik /* Number of sectors per track 1-255. Number of heads 1-16 */ 4663c6fd2807SJeff Garzik if (sectors < 1 || sectors > 255 || heads < 1 || heads > 16) 4664c6fd2807SJeff Garzik return AC_ERR_INVALID; 4665c6fd2807SJeff Garzik 4666c6fd2807SJeff Garzik /* set up init dev params taskfile */ 4667c6fd2807SJeff Garzik DPRINTK("init dev params \n"); 4668c6fd2807SJeff Garzik 4669c6fd2807SJeff Garzik ata_tf_init(dev, &tf); 4670c6fd2807SJeff Garzik tf.command = ATA_CMD_INIT_DEV_PARAMS; 4671c6fd2807SJeff Garzik tf.flags |= ATA_TFLAG_ISADDR | ATA_TFLAG_DEVICE; 4672c6fd2807SJeff Garzik tf.protocol = ATA_PROT_NODATA; 4673c6fd2807SJeff Garzik tf.nsect = sectors; 4674c6fd2807SJeff Garzik tf.device |= (heads - 1) & 0x0f; /* max head = num. of heads - 1 */ 4675c6fd2807SJeff Garzik 46762b789108STejun Heo err_mask = ata_exec_internal(dev, &tf, NULL, DMA_NONE, NULL, 0, 0); 467718b2466cSAlan Cox /* A clean abort indicates an original or just out of spec drive 467818b2466cSAlan Cox and we should continue as we issue the setup based on the 467918b2466cSAlan Cox drive reported working geometry */ 468018b2466cSAlan Cox if (err_mask == AC_ERR_DEV && (tf.feature & ATA_ABORTED)) 468118b2466cSAlan Cox err_mask = 0; 4682c6fd2807SJeff Garzik 4683c6fd2807SJeff Garzik DPRINTK("EXIT, err_mask=%x\n", err_mask); 4684c6fd2807SJeff Garzik return err_mask; 4685c6fd2807SJeff Garzik } 4686c6fd2807SJeff Garzik 4687c6fd2807SJeff Garzik /** 4688c6fd2807SJeff Garzik * ata_sg_clean - Unmap DMA memory associated with command 4689c6fd2807SJeff Garzik * @qc: Command containing DMA memory to be released 4690c6fd2807SJeff Garzik * 4691c6fd2807SJeff Garzik * Unmap all mapped DMA memory associated with this command. 4692c6fd2807SJeff Garzik * 4693c6fd2807SJeff Garzik * LOCKING: 4694cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 4695c6fd2807SJeff Garzik */ 469670e6ad0cSTejun Heo void ata_sg_clean(struct ata_queued_cmd *qc) 4697c6fd2807SJeff Garzik { 4698c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 4699ff2aeb1eSTejun Heo struct scatterlist *sg = qc->sg; 4700c6fd2807SJeff Garzik int dir = qc->dma_dir; 4701c6fd2807SJeff Garzik 4702c6fd2807SJeff Garzik WARN_ON(sg == NULL); 4703c6fd2807SJeff Garzik 4704dde20207SJames Bottomley VPRINTK("unmapping %u sg elements\n", qc->n_elem); 4705c6fd2807SJeff Garzik 4706dde20207SJames Bottomley if (qc->n_elem) 4707dde20207SJames Bottomley dma_unmap_sg(ap->dev, sg, qc->n_elem, dir); 4708c6fd2807SJeff Garzik 4709c6fd2807SJeff Garzik qc->flags &= ~ATA_QCFLAG_DMAMAP; 4710ff2aeb1eSTejun Heo qc->sg = NULL; 4711c6fd2807SJeff Garzik } 4712c6fd2807SJeff Garzik 4713c6fd2807SJeff Garzik /** 4714c6fd2807SJeff Garzik * ata_fill_sg - Fill PCI IDE PRD table 4715c6fd2807SJeff Garzik * @qc: Metadata associated with taskfile to be transferred 4716c6fd2807SJeff Garzik * 4717c6fd2807SJeff Garzik * Fill PCI IDE PRD (scatter-gather) table with segments 4718c6fd2807SJeff Garzik * associated with the current disk command. 4719c6fd2807SJeff Garzik * 4720c6fd2807SJeff Garzik * LOCKING: 4721cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 4722c6fd2807SJeff Garzik * 4723c6fd2807SJeff Garzik */ 4724c6fd2807SJeff Garzik static void ata_fill_sg(struct ata_queued_cmd *qc) 4725c6fd2807SJeff Garzik { 4726c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 4727c6fd2807SJeff Garzik struct scatterlist *sg; 4728ff2aeb1eSTejun Heo unsigned int si, pi; 4729c6fd2807SJeff Garzik 4730ff2aeb1eSTejun Heo pi = 0; 4731ff2aeb1eSTejun Heo for_each_sg(qc->sg, sg, qc->n_elem, si) { 4732c6fd2807SJeff Garzik u32 addr, offset; 4733c6fd2807SJeff Garzik u32 sg_len, len; 4734c6fd2807SJeff Garzik 4735c6fd2807SJeff Garzik /* determine if physical DMA addr spans 64K boundary. 4736c6fd2807SJeff Garzik * Note h/w doesn't support 64-bit, so we unconditionally 4737c6fd2807SJeff Garzik * truncate dma_addr_t to u32. 4738c6fd2807SJeff Garzik */ 4739c6fd2807SJeff Garzik addr = (u32) sg_dma_address(sg); 4740c6fd2807SJeff Garzik sg_len = sg_dma_len(sg); 4741c6fd2807SJeff Garzik 4742c6fd2807SJeff Garzik while (sg_len) { 4743c6fd2807SJeff Garzik offset = addr & 0xffff; 4744c6fd2807SJeff Garzik len = sg_len; 4745c6fd2807SJeff Garzik if ((offset + sg_len) > 0x10000) 4746c6fd2807SJeff Garzik len = 0x10000 - offset; 4747c6fd2807SJeff Garzik 4748ff2aeb1eSTejun Heo ap->prd[pi].addr = cpu_to_le32(addr); 4749ff2aeb1eSTejun Heo ap->prd[pi].flags_len = cpu_to_le32(len & 0xffff); 4750ff2aeb1eSTejun Heo VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", pi, addr, len); 4751c6fd2807SJeff Garzik 4752ff2aeb1eSTejun Heo pi++; 4753c6fd2807SJeff Garzik sg_len -= len; 4754c6fd2807SJeff Garzik addr += len; 4755c6fd2807SJeff Garzik } 4756c6fd2807SJeff Garzik } 4757c6fd2807SJeff Garzik 4758ff2aeb1eSTejun Heo ap->prd[pi - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); 4759c6fd2807SJeff Garzik } 4760b9a4197eSTejun Heo 4761c6fd2807SJeff Garzik /** 4762d26fc955SAlan Cox * ata_fill_sg_dumb - Fill PCI IDE PRD table 4763d26fc955SAlan Cox * @qc: Metadata associated with taskfile to be transferred 4764d26fc955SAlan Cox * 4765d26fc955SAlan Cox * Fill PCI IDE PRD (scatter-gather) table with segments 4766d26fc955SAlan Cox * associated with the current disk command. Perform the fill 4767d26fc955SAlan Cox * so that we avoid writing any length 64K records for 4768d26fc955SAlan Cox * controllers that don't follow the spec. 4769d26fc955SAlan Cox * 4770d26fc955SAlan Cox * LOCKING: 4771d26fc955SAlan Cox * spin_lock_irqsave(host lock) 4772d26fc955SAlan Cox * 4773d26fc955SAlan Cox */ 4774d26fc955SAlan Cox static void ata_fill_sg_dumb(struct ata_queued_cmd *qc) 4775d26fc955SAlan Cox { 4776d26fc955SAlan Cox struct ata_port *ap = qc->ap; 4777d26fc955SAlan Cox struct scatterlist *sg; 4778ff2aeb1eSTejun Heo unsigned int si, pi; 4779d26fc955SAlan Cox 4780ff2aeb1eSTejun Heo pi = 0; 4781ff2aeb1eSTejun Heo for_each_sg(qc->sg, sg, qc->n_elem, si) { 4782d26fc955SAlan Cox u32 addr, offset; 4783d26fc955SAlan Cox u32 sg_len, len, blen; 4784d26fc955SAlan Cox 4785d26fc955SAlan Cox /* determine if physical DMA addr spans 64K boundary. 4786d26fc955SAlan Cox * Note h/w doesn't support 64-bit, so we unconditionally 4787d26fc955SAlan Cox * truncate dma_addr_t to u32. 4788d26fc955SAlan Cox */ 4789d26fc955SAlan Cox addr = (u32) sg_dma_address(sg); 4790d26fc955SAlan Cox sg_len = sg_dma_len(sg); 4791d26fc955SAlan Cox 4792d26fc955SAlan Cox while (sg_len) { 4793d26fc955SAlan Cox offset = addr & 0xffff; 4794d26fc955SAlan Cox len = sg_len; 4795d26fc955SAlan Cox if ((offset + sg_len) > 0x10000) 4796d26fc955SAlan Cox len = 0x10000 - offset; 4797d26fc955SAlan Cox 4798d26fc955SAlan Cox blen = len & 0xffff; 4799ff2aeb1eSTejun Heo ap->prd[pi].addr = cpu_to_le32(addr); 4800d26fc955SAlan Cox if (blen == 0) { 4801d26fc955SAlan Cox /* Some PATA chipsets like the CS5530 can't 4802d26fc955SAlan Cox cope with 0x0000 meaning 64K as the spec says */ 4803ff2aeb1eSTejun Heo ap->prd[pi].flags_len = cpu_to_le32(0x8000); 4804d26fc955SAlan Cox blen = 0x8000; 4805ff2aeb1eSTejun Heo ap->prd[++pi].addr = cpu_to_le32(addr + 0x8000); 4806d26fc955SAlan Cox } 4807ff2aeb1eSTejun Heo ap->prd[pi].flags_len = cpu_to_le32(blen); 4808ff2aeb1eSTejun Heo VPRINTK("PRD[%u] = (0x%X, 0x%X)\n", pi, addr, len); 4809d26fc955SAlan Cox 4810ff2aeb1eSTejun Heo pi++; 4811d26fc955SAlan Cox sg_len -= len; 4812d26fc955SAlan Cox addr += len; 4813d26fc955SAlan Cox } 4814d26fc955SAlan Cox } 4815d26fc955SAlan Cox 4816ff2aeb1eSTejun Heo ap->prd[pi - 1].flags_len |= cpu_to_le32(ATA_PRD_EOT); 4817d26fc955SAlan Cox } 4818d26fc955SAlan Cox 4819d26fc955SAlan Cox /** 4820c6fd2807SJeff Garzik * ata_check_atapi_dma - Check whether ATAPI DMA can be supported 4821c6fd2807SJeff Garzik * @qc: Metadata associated with taskfile to check 4822c6fd2807SJeff Garzik * 4823c6fd2807SJeff Garzik * Allow low-level driver to filter ATA PACKET commands, returning 4824c6fd2807SJeff Garzik * a status indicating whether or not it is OK to use DMA for the 4825c6fd2807SJeff Garzik * supplied PACKET command. 4826c6fd2807SJeff Garzik * 4827c6fd2807SJeff Garzik * LOCKING: 4828cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 4829c6fd2807SJeff Garzik * 4830c6fd2807SJeff Garzik * RETURNS: 0 when ATAPI DMA can be used 4831c6fd2807SJeff Garzik * nonzero otherwise 4832c6fd2807SJeff Garzik */ 4833c6fd2807SJeff Garzik int ata_check_atapi_dma(struct ata_queued_cmd *qc) 4834c6fd2807SJeff Garzik { 4835c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 4836c6fd2807SJeff Garzik 4837b9a4197eSTejun Heo /* Don't allow DMA if it isn't multiple of 16 bytes. Quite a 4838b9a4197eSTejun Heo * few ATAPI devices choke on such DMA requests. 4839b9a4197eSTejun Heo */ 4840b9a4197eSTejun Heo if (unlikely(qc->nbytes & 15)) 48416f23a31dSAlbert Lee return 1; 48426f23a31dSAlbert Lee 4843c6fd2807SJeff Garzik if (ap->ops->check_atapi_dma) 4844b9a4197eSTejun Heo return ap->ops->check_atapi_dma(qc); 4845c6fd2807SJeff Garzik 4846b9a4197eSTejun Heo return 0; 4847c6fd2807SJeff Garzik } 4848b9a4197eSTejun Heo 4849c6fd2807SJeff Garzik /** 485031cc23b3STejun Heo * ata_std_qc_defer - Check whether a qc needs to be deferred 485131cc23b3STejun Heo * @qc: ATA command in question 485231cc23b3STejun Heo * 485331cc23b3STejun Heo * Non-NCQ commands cannot run with any other command, NCQ or 485431cc23b3STejun Heo * not. As upper layer only knows the queue depth, we are 485531cc23b3STejun Heo * responsible for maintaining exclusion. This function checks 485631cc23b3STejun Heo * whether a new command @qc can be issued. 485731cc23b3STejun Heo * 485831cc23b3STejun Heo * LOCKING: 485931cc23b3STejun Heo * spin_lock_irqsave(host lock) 486031cc23b3STejun Heo * 486131cc23b3STejun Heo * RETURNS: 486231cc23b3STejun Heo * ATA_DEFER_* if deferring is needed, 0 otherwise. 486331cc23b3STejun Heo */ 486431cc23b3STejun Heo int ata_std_qc_defer(struct ata_queued_cmd *qc) 486531cc23b3STejun Heo { 486631cc23b3STejun Heo struct ata_link *link = qc->dev->link; 486731cc23b3STejun Heo 486831cc23b3STejun Heo if (qc->tf.protocol == ATA_PROT_NCQ) { 486931cc23b3STejun Heo if (!ata_tag_valid(link->active_tag)) 487031cc23b3STejun Heo return 0; 487131cc23b3STejun Heo } else { 487231cc23b3STejun Heo if (!ata_tag_valid(link->active_tag) && !link->sactive) 487331cc23b3STejun Heo return 0; 487431cc23b3STejun Heo } 487531cc23b3STejun Heo 487631cc23b3STejun Heo return ATA_DEFER_LINK; 487731cc23b3STejun Heo } 487831cc23b3STejun Heo 487931cc23b3STejun Heo /** 4880c6fd2807SJeff Garzik * ata_qc_prep - Prepare taskfile for submission 4881c6fd2807SJeff Garzik * @qc: Metadata associated with taskfile to be prepared 4882c6fd2807SJeff Garzik * 4883c6fd2807SJeff Garzik * Prepare ATA taskfile for submission. 4884c6fd2807SJeff Garzik * 4885c6fd2807SJeff Garzik * LOCKING: 4886cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 4887c6fd2807SJeff Garzik */ 4888c6fd2807SJeff Garzik void ata_qc_prep(struct ata_queued_cmd *qc) 4889c6fd2807SJeff Garzik { 4890c6fd2807SJeff Garzik if (!(qc->flags & ATA_QCFLAG_DMAMAP)) 4891c6fd2807SJeff Garzik return; 4892c6fd2807SJeff Garzik 4893c6fd2807SJeff Garzik ata_fill_sg(qc); 4894c6fd2807SJeff Garzik } 4895c6fd2807SJeff Garzik 4896d26fc955SAlan Cox /** 4897d26fc955SAlan Cox * ata_dumb_qc_prep - Prepare taskfile for submission 4898d26fc955SAlan Cox * @qc: Metadata associated with taskfile to be prepared 4899d26fc955SAlan Cox * 4900d26fc955SAlan Cox * Prepare ATA taskfile for submission. 4901d26fc955SAlan Cox * 4902d26fc955SAlan Cox * LOCKING: 4903d26fc955SAlan Cox * spin_lock_irqsave(host lock) 4904d26fc955SAlan Cox */ 4905d26fc955SAlan Cox void ata_dumb_qc_prep(struct ata_queued_cmd *qc) 4906d26fc955SAlan Cox { 4907d26fc955SAlan Cox if (!(qc->flags & ATA_QCFLAG_DMAMAP)) 4908d26fc955SAlan Cox return; 4909d26fc955SAlan Cox 4910d26fc955SAlan Cox ata_fill_sg_dumb(qc); 4911d26fc955SAlan Cox } 4912d26fc955SAlan Cox 4913c6fd2807SJeff Garzik void ata_noop_qc_prep(struct ata_queued_cmd *qc) { } 4914c6fd2807SJeff Garzik 4915c6fd2807SJeff Garzik /** 4916c6fd2807SJeff Garzik * ata_sg_init - Associate command with scatter-gather table. 4917c6fd2807SJeff Garzik * @qc: Command to be associated 4918c6fd2807SJeff Garzik * @sg: Scatter-gather table. 4919c6fd2807SJeff Garzik * @n_elem: Number of elements in s/g table. 4920c6fd2807SJeff Garzik * 4921c6fd2807SJeff Garzik * Initialize the data-related elements of queued_cmd @qc 4922c6fd2807SJeff Garzik * to point to a scatter-gather table @sg, containing @n_elem 4923c6fd2807SJeff Garzik * elements. 4924c6fd2807SJeff Garzik * 4925c6fd2807SJeff Garzik * LOCKING: 4926cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 4927c6fd2807SJeff Garzik */ 4928c6fd2807SJeff Garzik void ata_sg_init(struct ata_queued_cmd *qc, struct scatterlist *sg, 4929c6fd2807SJeff Garzik unsigned int n_elem) 4930c6fd2807SJeff Garzik { 4931ff2aeb1eSTejun Heo qc->sg = sg; 4932c6fd2807SJeff Garzik qc->n_elem = n_elem; 4933ff2aeb1eSTejun Heo qc->cursg = qc->sg; 4934ff2aeb1eSTejun Heo } 4935ff2aeb1eSTejun Heo 4936c6fd2807SJeff Garzik /** 4937c6fd2807SJeff Garzik * ata_sg_setup - DMA-map the scatter-gather table associated with a command. 4938c6fd2807SJeff Garzik * @qc: Command with scatter-gather table to be mapped. 4939c6fd2807SJeff Garzik * 4940c6fd2807SJeff Garzik * DMA-map the scatter-gather table associated with queued_cmd @qc. 4941c6fd2807SJeff Garzik * 4942c6fd2807SJeff Garzik * LOCKING: 4943cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 4944c6fd2807SJeff Garzik * 4945c6fd2807SJeff Garzik * RETURNS: 4946c6fd2807SJeff Garzik * Zero on success, negative on error. 4947c6fd2807SJeff Garzik * 4948c6fd2807SJeff Garzik */ 4949c6fd2807SJeff Garzik static int ata_sg_setup(struct ata_queued_cmd *qc) 4950c6fd2807SJeff Garzik { 4951c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 4952dde20207SJames Bottomley unsigned int n_elem; 4953c6fd2807SJeff Garzik 495444877b4eSTejun Heo VPRINTK("ENTER, ata%u\n", ap->print_id); 4955c6fd2807SJeff Garzik 4956dde20207SJames Bottomley n_elem = dma_map_sg(ap->dev, qc->sg, qc->n_elem, qc->dma_dir); 4957dde20207SJames Bottomley if (n_elem < 1) 4958c6fd2807SJeff Garzik return -1; 4959c6fd2807SJeff Garzik 4960dde20207SJames Bottomley DPRINTK("%d sg elements mapped\n", n_elem); 4961dde20207SJames Bottomley 4962dde20207SJames Bottomley qc->n_elem = n_elem; 4963f92a2636STejun Heo qc->flags |= ATA_QCFLAG_DMAMAP; 4964c6fd2807SJeff Garzik 4965c6fd2807SJeff Garzik return 0; 4966c6fd2807SJeff Garzik } 4967c6fd2807SJeff Garzik 4968c6fd2807SJeff Garzik /** 4969c6fd2807SJeff Garzik * swap_buf_le16 - swap halves of 16-bit words in place 4970c6fd2807SJeff Garzik * @buf: Buffer to swap 4971c6fd2807SJeff Garzik * @buf_words: Number of 16-bit words in buffer. 4972c6fd2807SJeff Garzik * 4973c6fd2807SJeff Garzik * Swap halves of 16-bit words if needed to convert from 4974c6fd2807SJeff Garzik * little-endian byte order to native cpu byte order, or 4975c6fd2807SJeff Garzik * vice-versa. 4976c6fd2807SJeff Garzik * 4977c6fd2807SJeff Garzik * LOCKING: 4978c6fd2807SJeff Garzik * Inherited from caller. 4979c6fd2807SJeff Garzik */ 4980c6fd2807SJeff Garzik void swap_buf_le16(u16 *buf, unsigned int buf_words) 4981c6fd2807SJeff Garzik { 4982c6fd2807SJeff Garzik #ifdef __BIG_ENDIAN 4983c6fd2807SJeff Garzik unsigned int i; 4984c6fd2807SJeff Garzik 4985c6fd2807SJeff Garzik for (i = 0; i < buf_words; i++) 4986c6fd2807SJeff Garzik buf[i] = le16_to_cpu(buf[i]); 4987c6fd2807SJeff Garzik #endif /* __BIG_ENDIAN */ 4988c6fd2807SJeff Garzik } 4989c6fd2807SJeff Garzik 4990c6fd2807SJeff Garzik /** 49910d5ff566STejun Heo * ata_data_xfer - Transfer data by PIO 499255dba312STejun Heo * @dev: device to target 4993c6fd2807SJeff Garzik * @buf: data buffer 4994c6fd2807SJeff Garzik * @buflen: buffer length 49950affa456SLinus Nilsson * @rw: read/write 4996c6fd2807SJeff Garzik * 4997c6fd2807SJeff Garzik * Transfer data from/to the device data register by PIO. 4998c6fd2807SJeff Garzik * 4999c6fd2807SJeff Garzik * LOCKING: 5000c6fd2807SJeff Garzik * Inherited from caller. 500155dba312STejun Heo * 500255dba312STejun Heo * RETURNS: 500355dba312STejun Heo * Bytes consumed. 5004c6fd2807SJeff Garzik */ 500555dba312STejun Heo unsigned int ata_data_xfer(struct ata_device *dev, unsigned char *buf, 500655dba312STejun Heo unsigned int buflen, int rw) 5007c6fd2807SJeff Garzik { 500855dba312STejun Heo struct ata_port *ap = dev->link->ap; 500955dba312STejun Heo void __iomem *data_addr = ap->ioaddr.data_addr; 5010c6fd2807SJeff Garzik unsigned int words = buflen >> 1; 5011c6fd2807SJeff Garzik 5012c6fd2807SJeff Garzik /* Transfer multiple of 2 bytes */ 501355dba312STejun Heo if (rw == READ) 501455dba312STejun Heo ioread16_rep(data_addr, buf, words); 5015c6fd2807SJeff Garzik else 501655dba312STejun Heo iowrite16_rep(data_addr, buf, words); 5017c6fd2807SJeff Garzik 5018c6fd2807SJeff Garzik /* Transfer trailing 1 byte, if any. */ 5019c6fd2807SJeff Garzik if (unlikely(buflen & 0x01)) { 50204ca4e439SAl Viro __le16 align_buf[1] = { 0 }; 5021c6fd2807SJeff Garzik unsigned char *trailing_buf = buf + buflen - 1; 5022c6fd2807SJeff Garzik 502355dba312STejun Heo if (rw == READ) { 502455dba312STejun Heo align_buf[0] = cpu_to_le16(ioread16(data_addr)); 5025c6fd2807SJeff Garzik memcpy(trailing_buf, align_buf, 1); 502655dba312STejun Heo } else { 502755dba312STejun Heo memcpy(align_buf, trailing_buf, 1); 502855dba312STejun Heo iowrite16(le16_to_cpu(align_buf[0]), data_addr); 5029c6fd2807SJeff Garzik } 503055dba312STejun Heo words++; 5031c6fd2807SJeff Garzik } 503255dba312STejun Heo 503355dba312STejun Heo return words << 1; 5034c6fd2807SJeff Garzik } 5035c6fd2807SJeff Garzik 5036c6fd2807SJeff Garzik /** 50370d5ff566STejun Heo * ata_data_xfer_noirq - Transfer data by PIO 503855dba312STejun Heo * @dev: device to target 5039c6fd2807SJeff Garzik * @buf: data buffer 5040c6fd2807SJeff Garzik * @buflen: buffer length 50410affa456SLinus Nilsson * @rw: read/write 5042c6fd2807SJeff Garzik * 5043c6fd2807SJeff Garzik * Transfer data from/to the device data register by PIO. Do the 5044c6fd2807SJeff Garzik * transfer with interrupts disabled. 5045c6fd2807SJeff Garzik * 5046c6fd2807SJeff Garzik * LOCKING: 5047c6fd2807SJeff Garzik * Inherited from caller. 504855dba312STejun Heo * 504955dba312STejun Heo * RETURNS: 505055dba312STejun Heo * Bytes consumed. 5051c6fd2807SJeff Garzik */ 505255dba312STejun Heo unsigned int ata_data_xfer_noirq(struct ata_device *dev, unsigned char *buf, 505355dba312STejun Heo unsigned int buflen, int rw) 5054c6fd2807SJeff Garzik { 5055c6fd2807SJeff Garzik unsigned long flags; 505655dba312STejun Heo unsigned int consumed; 505755dba312STejun Heo 5058c6fd2807SJeff Garzik local_irq_save(flags); 505955dba312STejun Heo consumed = ata_data_xfer(dev, buf, buflen, rw); 5060c6fd2807SJeff Garzik local_irq_restore(flags); 506155dba312STejun Heo 506255dba312STejun Heo return consumed; 5063c6fd2807SJeff Garzik } 5064c6fd2807SJeff Garzik 5065c6fd2807SJeff Garzik 5066c6fd2807SJeff Garzik /** 50675a5dbd18SMark Lord * ata_pio_sector - Transfer a sector of data. 5068c6fd2807SJeff Garzik * @qc: Command on going 5069c6fd2807SJeff Garzik * 50705a5dbd18SMark Lord * Transfer qc->sect_size bytes of data from/to the ATA device. 5071c6fd2807SJeff Garzik * 5072c6fd2807SJeff Garzik * LOCKING: 5073c6fd2807SJeff Garzik * Inherited from caller. 5074c6fd2807SJeff Garzik */ 5075c6fd2807SJeff Garzik 5076c6fd2807SJeff Garzik static void ata_pio_sector(struct ata_queued_cmd *qc) 5077c6fd2807SJeff Garzik { 5078c6fd2807SJeff Garzik int do_write = (qc->tf.flags & ATA_TFLAG_WRITE); 5079c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 5080c6fd2807SJeff Garzik struct page *page; 5081c6fd2807SJeff Garzik unsigned int offset; 5082c6fd2807SJeff Garzik unsigned char *buf; 5083c6fd2807SJeff Garzik 50845a5dbd18SMark Lord if (qc->curbytes == qc->nbytes - qc->sect_size) 5085c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_LAST; 5086c6fd2807SJeff Garzik 508745711f1aSJens Axboe page = sg_page(qc->cursg); 508887260216SJens Axboe offset = qc->cursg->offset + qc->cursg_ofs; 5089c6fd2807SJeff Garzik 5090c6fd2807SJeff Garzik /* get the current page and offset */ 5091c6fd2807SJeff Garzik page = nth_page(page, (offset >> PAGE_SHIFT)); 5092c6fd2807SJeff Garzik offset %= PAGE_SIZE; 5093c6fd2807SJeff Garzik 5094c6fd2807SJeff Garzik DPRINTK("data %s\n", qc->tf.flags & ATA_TFLAG_WRITE ? "write" : "read"); 5095c6fd2807SJeff Garzik 5096c6fd2807SJeff Garzik if (PageHighMem(page)) { 5097c6fd2807SJeff Garzik unsigned long flags; 5098c6fd2807SJeff Garzik 5099c6fd2807SJeff Garzik /* FIXME: use a bounce buffer */ 5100c6fd2807SJeff Garzik local_irq_save(flags); 5101c6fd2807SJeff Garzik buf = kmap_atomic(page, KM_IRQ0); 5102c6fd2807SJeff Garzik 5103c6fd2807SJeff Garzik /* do the actual data transfer */ 51045a5dbd18SMark Lord ap->ops->data_xfer(qc->dev, buf + offset, qc->sect_size, do_write); 5105c6fd2807SJeff Garzik 5106c6fd2807SJeff Garzik kunmap_atomic(buf, KM_IRQ0); 5107c6fd2807SJeff Garzik local_irq_restore(flags); 5108c6fd2807SJeff Garzik } else { 5109c6fd2807SJeff Garzik buf = page_address(page); 51105a5dbd18SMark Lord ap->ops->data_xfer(qc->dev, buf + offset, qc->sect_size, do_write); 5111c6fd2807SJeff Garzik } 5112c6fd2807SJeff Garzik 51135a5dbd18SMark Lord qc->curbytes += qc->sect_size; 51145a5dbd18SMark Lord qc->cursg_ofs += qc->sect_size; 5115c6fd2807SJeff Garzik 511687260216SJens Axboe if (qc->cursg_ofs == qc->cursg->length) { 511787260216SJens Axboe qc->cursg = sg_next(qc->cursg); 5118c6fd2807SJeff Garzik qc->cursg_ofs = 0; 5119c6fd2807SJeff Garzik } 5120c6fd2807SJeff Garzik } 5121c6fd2807SJeff Garzik 5122c6fd2807SJeff Garzik /** 51235a5dbd18SMark Lord * ata_pio_sectors - Transfer one or many sectors. 5124c6fd2807SJeff Garzik * @qc: Command on going 5125c6fd2807SJeff Garzik * 51265a5dbd18SMark Lord * Transfer one or many sectors of data from/to the 5127c6fd2807SJeff Garzik * ATA device for the DRQ request. 5128c6fd2807SJeff Garzik * 5129c6fd2807SJeff Garzik * LOCKING: 5130c6fd2807SJeff Garzik * Inherited from caller. 5131c6fd2807SJeff Garzik */ 5132c6fd2807SJeff Garzik 5133c6fd2807SJeff Garzik static void ata_pio_sectors(struct ata_queued_cmd *qc) 5134c6fd2807SJeff Garzik { 5135c6fd2807SJeff Garzik if (is_multi_taskfile(&qc->tf)) { 5136c6fd2807SJeff Garzik /* READ/WRITE MULTIPLE */ 5137c6fd2807SJeff Garzik unsigned int nsect; 5138c6fd2807SJeff Garzik 5139c6fd2807SJeff Garzik WARN_ON(qc->dev->multi_count == 0); 5140c6fd2807SJeff Garzik 51415a5dbd18SMark Lord nsect = min((qc->nbytes - qc->curbytes) / qc->sect_size, 5142726f0785STejun Heo qc->dev->multi_count); 5143c6fd2807SJeff Garzik while (nsect--) 5144c6fd2807SJeff Garzik ata_pio_sector(qc); 5145c6fd2807SJeff Garzik } else 5146c6fd2807SJeff Garzik ata_pio_sector(qc); 51474cc980b3SAlbert Lee 51484cc980b3SAlbert Lee ata_altstatus(qc->ap); /* flush */ 5149c6fd2807SJeff Garzik } 5150c6fd2807SJeff Garzik 5151c6fd2807SJeff Garzik /** 5152c6fd2807SJeff Garzik * atapi_send_cdb - Write CDB bytes to hardware 5153c6fd2807SJeff Garzik * @ap: Port to which ATAPI device is attached. 5154c6fd2807SJeff Garzik * @qc: Taskfile currently active 5155c6fd2807SJeff Garzik * 5156c6fd2807SJeff Garzik * When device has indicated its readiness to accept 5157c6fd2807SJeff Garzik * a CDB, this function is called. Send the CDB. 5158c6fd2807SJeff Garzik * 5159c6fd2807SJeff Garzik * LOCKING: 5160c6fd2807SJeff Garzik * caller. 5161c6fd2807SJeff Garzik */ 5162c6fd2807SJeff Garzik 5163c6fd2807SJeff Garzik static void atapi_send_cdb(struct ata_port *ap, struct ata_queued_cmd *qc) 5164c6fd2807SJeff Garzik { 5165c6fd2807SJeff Garzik /* send SCSI cdb */ 5166c6fd2807SJeff Garzik DPRINTK("send cdb\n"); 5167c6fd2807SJeff Garzik WARN_ON(qc->dev->cdb_len < 12); 5168c6fd2807SJeff Garzik 5169c6fd2807SJeff Garzik ap->ops->data_xfer(qc->dev, qc->cdb, qc->dev->cdb_len, 1); 5170c6fd2807SJeff Garzik ata_altstatus(ap); /* flush */ 5171c6fd2807SJeff Garzik 5172c6fd2807SJeff Garzik switch (qc->tf.protocol) { 51730dc36888STejun Heo case ATAPI_PROT_PIO: 5174c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST; 5175c6fd2807SJeff Garzik break; 51760dc36888STejun Heo case ATAPI_PROT_NODATA: 5177c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_LAST; 5178c6fd2807SJeff Garzik break; 51790dc36888STejun Heo case ATAPI_PROT_DMA: 5180c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_LAST; 5181c6fd2807SJeff Garzik /* initiate bmdma */ 5182c6fd2807SJeff Garzik ap->ops->bmdma_start(qc); 5183c6fd2807SJeff Garzik break; 5184c6fd2807SJeff Garzik } 5185c6fd2807SJeff Garzik } 5186c6fd2807SJeff Garzik 5187c6fd2807SJeff Garzik /** 5188c6fd2807SJeff Garzik * __atapi_pio_bytes - Transfer data from/to the ATAPI device. 5189c6fd2807SJeff Garzik * @qc: Command on going 5190c6fd2807SJeff Garzik * @bytes: number of bytes 5191c6fd2807SJeff Garzik * 5192c6fd2807SJeff Garzik * Transfer Transfer data from/to the ATAPI device. 5193c6fd2807SJeff Garzik * 5194c6fd2807SJeff Garzik * LOCKING: 5195c6fd2807SJeff Garzik * Inherited from caller. 5196c6fd2807SJeff Garzik * 5197c6fd2807SJeff Garzik */ 5198140b5e59STejun Heo static int __atapi_pio_bytes(struct ata_queued_cmd *qc, unsigned int bytes) 5199c6fd2807SJeff Garzik { 520056c819dfSTejun Heo int rw = (qc->tf.flags & ATA_TFLAG_WRITE) ? WRITE : READ; 5201c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 520256c819dfSTejun Heo struct ata_device *dev = qc->dev; 520356c819dfSTejun Heo struct ata_eh_info *ehi = &dev->link->eh_info; 5204140b5e59STejun Heo struct scatterlist *sg; 5205c6fd2807SJeff Garzik struct page *page; 5206c6fd2807SJeff Garzik unsigned char *buf; 520756c819dfSTejun Heo unsigned int offset, count, consumed; 5208c6fd2807SJeff Garzik 5209c6fd2807SJeff Garzik next_sg: 5210140b5e59STejun Heo sg = qc->cursg; 5211140b5e59STejun Heo if (unlikely(!sg)) { 5212fa2fc7f4SJames Bottomley ata_ehi_push_desc(ehi, "unexpected or too much trailing data " 5213140b5e59STejun Heo "buf=%u cur=%u bytes=%u", 5214140b5e59STejun Heo qc->nbytes, qc->curbytes, bytes); 5215140b5e59STejun Heo return -1; 5216c6fd2807SJeff Garzik } 5217c6fd2807SJeff Garzik 521845711f1aSJens Axboe page = sg_page(sg); 5219c6fd2807SJeff Garzik offset = sg->offset + qc->cursg_ofs; 5220c6fd2807SJeff Garzik 5221c6fd2807SJeff Garzik /* get the current page and offset */ 5222c6fd2807SJeff Garzik page = nth_page(page, (offset >> PAGE_SHIFT)); 5223c6fd2807SJeff Garzik offset %= PAGE_SIZE; 5224c6fd2807SJeff Garzik 5225c6fd2807SJeff Garzik /* don't overrun current sg */ 5226c6fd2807SJeff Garzik count = min(sg->length - qc->cursg_ofs, bytes); 5227c6fd2807SJeff Garzik 5228c6fd2807SJeff Garzik /* don't cross page boundaries */ 5229c6fd2807SJeff Garzik count = min(count, (unsigned int)PAGE_SIZE - offset); 5230c6fd2807SJeff Garzik 5231c6fd2807SJeff Garzik DPRINTK("data %s\n", qc->tf.flags & ATA_TFLAG_WRITE ? "write" : "read"); 5232c6fd2807SJeff Garzik 5233c6fd2807SJeff Garzik if (PageHighMem(page)) { 5234c6fd2807SJeff Garzik unsigned long flags; 5235c6fd2807SJeff Garzik 5236c6fd2807SJeff Garzik /* FIXME: use bounce buffer */ 5237c6fd2807SJeff Garzik local_irq_save(flags); 5238c6fd2807SJeff Garzik buf = kmap_atomic(page, KM_IRQ0); 5239c6fd2807SJeff Garzik 5240c6fd2807SJeff Garzik /* do the actual data transfer */ 524156c819dfSTejun Heo consumed = ap->ops->data_xfer(dev, buf + offset, count, rw); 5242c6fd2807SJeff Garzik 5243c6fd2807SJeff Garzik kunmap_atomic(buf, KM_IRQ0); 5244c6fd2807SJeff Garzik local_irq_restore(flags); 5245c6fd2807SJeff Garzik } else { 5246c6fd2807SJeff Garzik buf = page_address(page); 524756c819dfSTejun Heo consumed = ap->ops->data_xfer(dev, buf + offset, count, rw); 5248c6fd2807SJeff Garzik } 5249c6fd2807SJeff Garzik 525056c819dfSTejun Heo bytes -= min(bytes, consumed); 5251c6fd2807SJeff Garzik qc->curbytes += count; 5252c6fd2807SJeff Garzik qc->cursg_ofs += count; 5253c6fd2807SJeff Garzik 5254c6fd2807SJeff Garzik if (qc->cursg_ofs == sg->length) { 525587260216SJens Axboe qc->cursg = sg_next(qc->cursg); 5256c6fd2807SJeff Garzik qc->cursg_ofs = 0; 5257c6fd2807SJeff Garzik } 5258c6fd2807SJeff Garzik 525956c819dfSTejun Heo /* consumed can be larger than count only for the last transfer */ 526056c819dfSTejun Heo WARN_ON(qc->cursg && count != consumed); 526156c819dfSTejun Heo 5262c6fd2807SJeff Garzik if (bytes) 5263c6fd2807SJeff Garzik goto next_sg; 5264140b5e59STejun Heo return 0; 5265c6fd2807SJeff Garzik } 5266c6fd2807SJeff Garzik 5267c6fd2807SJeff Garzik /** 5268c6fd2807SJeff Garzik * atapi_pio_bytes - Transfer data from/to the ATAPI device. 5269c6fd2807SJeff Garzik * @qc: Command on going 5270c6fd2807SJeff Garzik * 5271c6fd2807SJeff Garzik * Transfer Transfer data from/to the ATAPI device. 5272c6fd2807SJeff Garzik * 5273c6fd2807SJeff Garzik * LOCKING: 5274c6fd2807SJeff Garzik * Inherited from caller. 5275c6fd2807SJeff Garzik */ 5276c6fd2807SJeff Garzik 5277c6fd2807SJeff Garzik static void atapi_pio_bytes(struct ata_queued_cmd *qc) 5278c6fd2807SJeff Garzik { 5279c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 5280c6fd2807SJeff Garzik struct ata_device *dev = qc->dev; 528156c819dfSTejun Heo struct ata_eh_info *ehi = &dev->link->eh_info; 5282c6fd2807SJeff Garzik unsigned int ireason, bc_lo, bc_hi, bytes; 5283c6fd2807SJeff Garzik int i_write, do_write = (qc->tf.flags & ATA_TFLAG_WRITE) ? 1 : 0; 5284c6fd2807SJeff Garzik 5285c6fd2807SJeff Garzik /* Abuse qc->result_tf for temp storage of intermediate TF 5286c6fd2807SJeff Garzik * here to save some kernel stack usage. 5287c6fd2807SJeff Garzik * For normal completion, qc->result_tf is not relevant. For 5288c6fd2807SJeff Garzik * error, qc->result_tf is later overwritten by ata_qc_complete(). 5289c6fd2807SJeff Garzik * So, the correctness of qc->result_tf is not affected. 5290c6fd2807SJeff Garzik */ 5291c6fd2807SJeff Garzik ap->ops->tf_read(ap, &qc->result_tf); 5292c6fd2807SJeff Garzik ireason = qc->result_tf.nsect; 5293c6fd2807SJeff Garzik bc_lo = qc->result_tf.lbam; 5294c6fd2807SJeff Garzik bc_hi = qc->result_tf.lbah; 5295c6fd2807SJeff Garzik bytes = (bc_hi << 8) | bc_lo; 5296c6fd2807SJeff Garzik 5297c6fd2807SJeff Garzik /* shall be cleared to zero, indicating xfer of data */ 52980106372dSAlbert Lee if (unlikely(ireason & (1 << 0))) 529956c819dfSTejun Heo goto atapi_check; 5300c6fd2807SJeff Garzik 5301c6fd2807SJeff Garzik /* make sure transfer direction matches expected */ 5302c6fd2807SJeff Garzik i_write = ((ireason & (1 << 1)) == 0) ? 1 : 0; 53030106372dSAlbert Lee if (unlikely(do_write != i_write)) 530456c819dfSTejun Heo goto atapi_check; 53050106372dSAlbert Lee 53060106372dSAlbert Lee if (unlikely(!bytes)) 530756c819dfSTejun Heo goto atapi_check; 5308c6fd2807SJeff Garzik 530944877b4eSTejun Heo VPRINTK("ata%u: xfering %d bytes\n", ap->print_id, bytes); 5310c6fd2807SJeff Garzik 531156c819dfSTejun Heo if (unlikely(__atapi_pio_bytes(qc, bytes))) 5312140b5e59STejun Heo goto err_out; 53134cc980b3SAlbert Lee ata_altstatus(ap); /* flush */ 5314c6fd2807SJeff Garzik 5315c6fd2807SJeff Garzik return; 5316c6fd2807SJeff Garzik 531756c819dfSTejun Heo atapi_check: 531856c819dfSTejun Heo ata_ehi_push_desc(ehi, "ATAPI check failed (ireason=0x%x bytes=%u)", 531956c819dfSTejun Heo ireason, bytes); 5320c6fd2807SJeff Garzik err_out: 5321c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_HSM; 5322c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 5323c6fd2807SJeff Garzik } 5324c6fd2807SJeff Garzik 5325c6fd2807SJeff Garzik /** 5326c6fd2807SJeff Garzik * ata_hsm_ok_in_wq - Check if the qc can be handled in the workqueue. 5327c6fd2807SJeff Garzik * @ap: the target ata_port 5328c6fd2807SJeff Garzik * @qc: qc on going 5329c6fd2807SJeff Garzik * 5330c6fd2807SJeff Garzik * RETURNS: 5331c6fd2807SJeff Garzik * 1 if ok in workqueue, 0 otherwise. 5332c6fd2807SJeff Garzik */ 5333c6fd2807SJeff Garzik 5334c6fd2807SJeff Garzik static inline int ata_hsm_ok_in_wq(struct ata_port *ap, struct ata_queued_cmd *qc) 5335c6fd2807SJeff Garzik { 5336c6fd2807SJeff Garzik if (qc->tf.flags & ATA_TFLAG_POLLING) 5337c6fd2807SJeff Garzik return 1; 5338c6fd2807SJeff Garzik 5339c6fd2807SJeff Garzik if (ap->hsm_task_state == HSM_ST_FIRST) { 5340c6fd2807SJeff Garzik if (qc->tf.protocol == ATA_PROT_PIO && 5341c6fd2807SJeff Garzik (qc->tf.flags & ATA_TFLAG_WRITE)) 5342c6fd2807SJeff Garzik return 1; 5343c6fd2807SJeff Garzik 5344405e66b3STejun Heo if (ata_is_atapi(qc->tf.protocol) && 5345c6fd2807SJeff Garzik !(qc->dev->flags & ATA_DFLAG_CDB_INTR)) 5346c6fd2807SJeff Garzik return 1; 5347c6fd2807SJeff Garzik } 5348c6fd2807SJeff Garzik 5349c6fd2807SJeff Garzik return 0; 5350c6fd2807SJeff Garzik } 5351c6fd2807SJeff Garzik 5352c6fd2807SJeff Garzik /** 5353c6fd2807SJeff Garzik * ata_hsm_qc_complete - finish a qc running on standard HSM 5354c6fd2807SJeff Garzik * @qc: Command to complete 5355c6fd2807SJeff Garzik * @in_wq: 1 if called from workqueue, 0 otherwise 5356c6fd2807SJeff Garzik * 5357c6fd2807SJeff Garzik * Finish @qc which is running on standard HSM. 5358c6fd2807SJeff Garzik * 5359c6fd2807SJeff Garzik * LOCKING: 5360cca3974eSJeff Garzik * If @in_wq is zero, spin_lock_irqsave(host lock). 5361c6fd2807SJeff Garzik * Otherwise, none on entry and grabs host lock. 5362c6fd2807SJeff Garzik */ 5363c6fd2807SJeff Garzik static void ata_hsm_qc_complete(struct ata_queued_cmd *qc, int in_wq) 5364c6fd2807SJeff Garzik { 5365c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 5366c6fd2807SJeff Garzik unsigned long flags; 5367c6fd2807SJeff Garzik 5368c6fd2807SJeff Garzik if (ap->ops->error_handler) { 5369c6fd2807SJeff Garzik if (in_wq) { 5370c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 5371c6fd2807SJeff Garzik 5372cca3974eSJeff Garzik /* EH might have kicked in while host lock is 5373cca3974eSJeff Garzik * released. 5374c6fd2807SJeff Garzik */ 5375c6fd2807SJeff Garzik qc = ata_qc_from_tag(ap, qc->tag); 5376c6fd2807SJeff Garzik if (qc) { 5377c6fd2807SJeff Garzik if (likely(!(qc->err_mask & AC_ERR_HSM))) { 537883625006SAkira Iguchi ap->ops->irq_on(ap); 5379c6fd2807SJeff Garzik ata_qc_complete(qc); 5380c6fd2807SJeff Garzik } else 5381c6fd2807SJeff Garzik ata_port_freeze(ap); 5382c6fd2807SJeff Garzik } 5383c6fd2807SJeff Garzik 5384c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 5385c6fd2807SJeff Garzik } else { 5386c6fd2807SJeff Garzik if (likely(!(qc->err_mask & AC_ERR_HSM))) 5387c6fd2807SJeff Garzik ata_qc_complete(qc); 5388c6fd2807SJeff Garzik else 5389c6fd2807SJeff Garzik ata_port_freeze(ap); 5390c6fd2807SJeff Garzik } 5391c6fd2807SJeff Garzik } else { 5392c6fd2807SJeff Garzik if (in_wq) { 5393c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 539483625006SAkira Iguchi ap->ops->irq_on(ap); 5395c6fd2807SJeff Garzik ata_qc_complete(qc); 5396c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 5397c6fd2807SJeff Garzik } else 5398c6fd2807SJeff Garzik ata_qc_complete(qc); 5399c6fd2807SJeff Garzik } 5400c6fd2807SJeff Garzik } 5401c6fd2807SJeff Garzik 5402c6fd2807SJeff Garzik /** 5403c6fd2807SJeff Garzik * ata_hsm_move - move the HSM to the next state. 5404c6fd2807SJeff Garzik * @ap: the target ata_port 5405c6fd2807SJeff Garzik * @qc: qc on going 5406c6fd2807SJeff Garzik * @status: current device status 5407c6fd2807SJeff Garzik * @in_wq: 1 if called from workqueue, 0 otherwise 5408c6fd2807SJeff Garzik * 5409c6fd2807SJeff Garzik * RETURNS: 5410c6fd2807SJeff Garzik * 1 when poll next status needed, 0 otherwise. 5411c6fd2807SJeff Garzik */ 5412c6fd2807SJeff Garzik int ata_hsm_move(struct ata_port *ap, struct ata_queued_cmd *qc, 5413c6fd2807SJeff Garzik u8 status, int in_wq) 5414c6fd2807SJeff Garzik { 5415c6fd2807SJeff Garzik unsigned long flags = 0; 5416c6fd2807SJeff Garzik int poll_next; 5417c6fd2807SJeff Garzik 5418c6fd2807SJeff Garzik WARN_ON((qc->flags & ATA_QCFLAG_ACTIVE) == 0); 5419c6fd2807SJeff Garzik 5420c6fd2807SJeff Garzik /* Make sure ata_qc_issue_prot() does not throw things 5421c6fd2807SJeff Garzik * like DMA polling into the workqueue. Notice that 5422c6fd2807SJeff Garzik * in_wq is not equivalent to (qc->tf.flags & ATA_TFLAG_POLLING). 5423c6fd2807SJeff Garzik */ 5424c6fd2807SJeff Garzik WARN_ON(in_wq != ata_hsm_ok_in_wq(ap, qc)); 5425c6fd2807SJeff Garzik 5426c6fd2807SJeff Garzik fsm_start: 5427c6fd2807SJeff Garzik DPRINTK("ata%u: protocol %d task_state %d (dev_stat 0x%X)\n", 542844877b4eSTejun Heo ap->print_id, qc->tf.protocol, ap->hsm_task_state, status); 5429c6fd2807SJeff Garzik 5430c6fd2807SJeff Garzik switch (ap->hsm_task_state) { 5431c6fd2807SJeff Garzik case HSM_ST_FIRST: 5432c6fd2807SJeff Garzik /* Send first data block or PACKET CDB */ 5433c6fd2807SJeff Garzik 5434c6fd2807SJeff Garzik /* If polling, we will stay in the work queue after 5435c6fd2807SJeff Garzik * sending the data. Otherwise, interrupt handler 5436c6fd2807SJeff Garzik * takes over after sending the data. 5437c6fd2807SJeff Garzik */ 5438c6fd2807SJeff Garzik poll_next = (qc->tf.flags & ATA_TFLAG_POLLING); 5439c6fd2807SJeff Garzik 5440c6fd2807SJeff Garzik /* check device status */ 5441c6fd2807SJeff Garzik if (unlikely((status & ATA_DRQ) == 0)) { 5442c6fd2807SJeff Garzik /* handle BSY=0, DRQ=0 as error */ 5443c6fd2807SJeff Garzik if (likely(status & (ATA_ERR | ATA_DF))) 5444c6fd2807SJeff Garzik /* device stops HSM for abort/error */ 5445c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_DEV; 5446c6fd2807SJeff Garzik else 5447c6fd2807SJeff Garzik /* HSM violation. Let EH handle this */ 5448c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_HSM; 5449c6fd2807SJeff Garzik 5450c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 5451c6fd2807SJeff Garzik goto fsm_start; 5452c6fd2807SJeff Garzik } 5453c6fd2807SJeff Garzik 5454c6fd2807SJeff Garzik /* Device should not ask for data transfer (DRQ=1) 5455c6fd2807SJeff Garzik * when it finds something wrong. 5456c6fd2807SJeff Garzik * We ignore DRQ here and stop the HSM by 5457c6fd2807SJeff Garzik * changing hsm_task_state to HSM_ST_ERR and 5458c6fd2807SJeff Garzik * let the EH abort the command or reset the device. 5459c6fd2807SJeff Garzik */ 5460c6fd2807SJeff Garzik if (unlikely(status & (ATA_ERR | ATA_DF))) { 54612d3b8eeaSAlbert Lee /* Some ATAPI tape drives forget to clear the ERR bit 54622d3b8eeaSAlbert Lee * when doing the next command (mostly request sense). 54632d3b8eeaSAlbert Lee * We ignore ERR here to workaround and proceed sending 54642d3b8eeaSAlbert Lee * the CDB. 54652d3b8eeaSAlbert Lee */ 54662d3b8eeaSAlbert Lee if (!(qc->dev->horkage & ATA_HORKAGE_STUCK_ERR)) { 54672d3b8eeaSAlbert Lee ata_port_printk(ap, KERN_WARNING, 54682d3b8eeaSAlbert Lee "DRQ=1 with device error, " 54692d3b8eeaSAlbert Lee "dev_stat 0x%X\n", status); 5470c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_HSM; 5471c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 5472c6fd2807SJeff Garzik goto fsm_start; 5473c6fd2807SJeff Garzik } 54742d3b8eeaSAlbert Lee } 5475c6fd2807SJeff Garzik 5476c6fd2807SJeff Garzik /* Send the CDB (atapi) or the first data block (ata pio out). 5477c6fd2807SJeff Garzik * During the state transition, interrupt handler shouldn't 5478c6fd2807SJeff Garzik * be invoked before the data transfer is complete and 5479c6fd2807SJeff Garzik * hsm_task_state is changed. Hence, the following locking. 5480c6fd2807SJeff Garzik */ 5481c6fd2807SJeff Garzik if (in_wq) 5482c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 5483c6fd2807SJeff Garzik 5484c6fd2807SJeff Garzik if (qc->tf.protocol == ATA_PROT_PIO) { 5485c6fd2807SJeff Garzik /* PIO data out protocol. 5486c6fd2807SJeff Garzik * send first data block. 5487c6fd2807SJeff Garzik */ 5488c6fd2807SJeff Garzik 5489c6fd2807SJeff Garzik /* ata_pio_sectors() might change the state 5490c6fd2807SJeff Garzik * to HSM_ST_LAST. so, the state is changed here 5491c6fd2807SJeff Garzik * before ata_pio_sectors(). 5492c6fd2807SJeff Garzik */ 5493c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST; 5494c6fd2807SJeff Garzik ata_pio_sectors(qc); 5495c6fd2807SJeff Garzik } else 5496c6fd2807SJeff Garzik /* send CDB */ 5497c6fd2807SJeff Garzik atapi_send_cdb(ap, qc); 5498c6fd2807SJeff Garzik 5499c6fd2807SJeff Garzik if (in_wq) 5500c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 5501c6fd2807SJeff Garzik 5502c6fd2807SJeff Garzik /* if polling, ata_pio_task() handles the rest. 5503c6fd2807SJeff Garzik * otherwise, interrupt handler takes over from here. 5504c6fd2807SJeff Garzik */ 5505c6fd2807SJeff Garzik break; 5506c6fd2807SJeff Garzik 5507c6fd2807SJeff Garzik case HSM_ST: 5508c6fd2807SJeff Garzik /* complete command or read/write the data register */ 55090dc36888STejun Heo if (qc->tf.protocol == ATAPI_PROT_PIO) { 5510c6fd2807SJeff Garzik /* ATAPI PIO protocol */ 5511c6fd2807SJeff Garzik if ((status & ATA_DRQ) == 0) { 5512c6fd2807SJeff Garzik /* No more data to transfer or device error. 5513c6fd2807SJeff Garzik * Device error will be tagged in HSM_ST_LAST. 5514c6fd2807SJeff Garzik */ 5515c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_LAST; 5516c6fd2807SJeff Garzik goto fsm_start; 5517c6fd2807SJeff Garzik } 5518c6fd2807SJeff Garzik 5519c6fd2807SJeff Garzik /* Device should not ask for data transfer (DRQ=1) 5520c6fd2807SJeff Garzik * when it finds something wrong. 5521c6fd2807SJeff Garzik * We ignore DRQ here and stop the HSM by 5522c6fd2807SJeff Garzik * changing hsm_task_state to HSM_ST_ERR and 5523c6fd2807SJeff Garzik * let the EH abort the command or reset the device. 5524c6fd2807SJeff Garzik */ 5525c6fd2807SJeff Garzik if (unlikely(status & (ATA_ERR | ATA_DF))) { 552644877b4eSTejun Heo ata_port_printk(ap, KERN_WARNING, "DRQ=1 with " 552744877b4eSTejun Heo "device error, dev_stat 0x%X\n", 552844877b4eSTejun Heo status); 5529c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_HSM; 5530c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 5531c6fd2807SJeff Garzik goto fsm_start; 5532c6fd2807SJeff Garzik } 5533c6fd2807SJeff Garzik 5534c6fd2807SJeff Garzik atapi_pio_bytes(qc); 5535c6fd2807SJeff Garzik 5536c6fd2807SJeff Garzik if (unlikely(ap->hsm_task_state == HSM_ST_ERR)) 5537c6fd2807SJeff Garzik /* bad ireason reported by device */ 5538c6fd2807SJeff Garzik goto fsm_start; 5539c6fd2807SJeff Garzik 5540c6fd2807SJeff Garzik } else { 5541c6fd2807SJeff Garzik /* ATA PIO protocol */ 5542c6fd2807SJeff Garzik if (unlikely((status & ATA_DRQ) == 0)) { 5543c6fd2807SJeff Garzik /* handle BSY=0, DRQ=0 as error */ 5544c6fd2807SJeff Garzik if (likely(status & (ATA_ERR | ATA_DF))) 5545c6fd2807SJeff Garzik /* device stops HSM for abort/error */ 5546c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_DEV; 5547c6fd2807SJeff Garzik else 554855a8e2c8STejun Heo /* HSM violation. Let EH handle this. 554955a8e2c8STejun Heo * Phantom devices also trigger this 555055a8e2c8STejun Heo * condition. Mark hint. 555155a8e2c8STejun Heo */ 555255a8e2c8STejun Heo qc->err_mask |= AC_ERR_HSM | 555355a8e2c8STejun Heo AC_ERR_NODEV_HINT; 5554c6fd2807SJeff Garzik 5555c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 5556c6fd2807SJeff Garzik goto fsm_start; 5557c6fd2807SJeff Garzik } 5558c6fd2807SJeff Garzik 5559c6fd2807SJeff Garzik /* For PIO reads, some devices may ask for 5560c6fd2807SJeff Garzik * data transfer (DRQ=1) alone with ERR=1. 5561c6fd2807SJeff Garzik * We respect DRQ here and transfer one 5562c6fd2807SJeff Garzik * block of junk data before changing the 5563c6fd2807SJeff Garzik * hsm_task_state to HSM_ST_ERR. 5564c6fd2807SJeff Garzik * 5565c6fd2807SJeff Garzik * For PIO writes, ERR=1 DRQ=1 doesn't make 5566c6fd2807SJeff Garzik * sense since the data block has been 5567c6fd2807SJeff Garzik * transferred to the device. 5568c6fd2807SJeff Garzik */ 5569c6fd2807SJeff Garzik if (unlikely(status & (ATA_ERR | ATA_DF))) { 5570c6fd2807SJeff Garzik /* data might be corrputed */ 5571c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_DEV; 5572c6fd2807SJeff Garzik 5573c6fd2807SJeff Garzik if (!(qc->tf.flags & ATA_TFLAG_WRITE)) { 5574c6fd2807SJeff Garzik ata_pio_sectors(qc); 5575c6fd2807SJeff Garzik status = ata_wait_idle(ap); 5576c6fd2807SJeff Garzik } 5577c6fd2807SJeff Garzik 5578c6fd2807SJeff Garzik if (status & (ATA_BUSY | ATA_DRQ)) 5579c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_HSM; 5580c6fd2807SJeff Garzik 5581c6fd2807SJeff Garzik /* ata_pio_sectors() might change the 5582c6fd2807SJeff Garzik * state to HSM_ST_LAST. so, the state 5583c6fd2807SJeff Garzik * is changed after ata_pio_sectors(). 5584c6fd2807SJeff Garzik */ 5585c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 5586c6fd2807SJeff Garzik goto fsm_start; 5587c6fd2807SJeff Garzik } 5588c6fd2807SJeff Garzik 5589c6fd2807SJeff Garzik ata_pio_sectors(qc); 5590c6fd2807SJeff Garzik 5591c6fd2807SJeff Garzik if (ap->hsm_task_state == HSM_ST_LAST && 5592c6fd2807SJeff Garzik (!(qc->tf.flags & ATA_TFLAG_WRITE))) { 5593c6fd2807SJeff Garzik /* all data read */ 5594c6fd2807SJeff Garzik status = ata_wait_idle(ap); 5595c6fd2807SJeff Garzik goto fsm_start; 5596c6fd2807SJeff Garzik } 5597c6fd2807SJeff Garzik } 5598c6fd2807SJeff Garzik 5599c6fd2807SJeff Garzik poll_next = 1; 5600c6fd2807SJeff Garzik break; 5601c6fd2807SJeff Garzik 5602c6fd2807SJeff Garzik case HSM_ST_LAST: 5603c6fd2807SJeff Garzik if (unlikely(!ata_ok(status))) { 5604c6fd2807SJeff Garzik qc->err_mask |= __ac_err_mask(status); 5605c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 5606c6fd2807SJeff Garzik goto fsm_start; 5607c6fd2807SJeff Garzik } 5608c6fd2807SJeff Garzik 5609c6fd2807SJeff Garzik /* no more data to transfer */ 5610c6fd2807SJeff Garzik DPRINTK("ata%u: dev %u command complete, drv_stat 0x%x\n", 561144877b4eSTejun Heo ap->print_id, qc->dev->devno, status); 5612c6fd2807SJeff Garzik 5613c6fd2807SJeff Garzik WARN_ON(qc->err_mask); 5614c6fd2807SJeff Garzik 5615c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_IDLE; 5616c6fd2807SJeff Garzik 5617c6fd2807SJeff Garzik /* complete taskfile transaction */ 5618c6fd2807SJeff Garzik ata_hsm_qc_complete(qc, in_wq); 5619c6fd2807SJeff Garzik 5620c6fd2807SJeff Garzik poll_next = 0; 5621c6fd2807SJeff Garzik break; 5622c6fd2807SJeff Garzik 5623c6fd2807SJeff Garzik case HSM_ST_ERR: 5624c6fd2807SJeff Garzik /* make sure qc->err_mask is available to 5625c6fd2807SJeff Garzik * know what's wrong and recover 5626c6fd2807SJeff Garzik */ 5627c6fd2807SJeff Garzik WARN_ON(qc->err_mask == 0); 5628c6fd2807SJeff Garzik 5629c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_IDLE; 5630c6fd2807SJeff Garzik 5631c6fd2807SJeff Garzik /* complete taskfile transaction */ 5632c6fd2807SJeff Garzik ata_hsm_qc_complete(qc, in_wq); 5633c6fd2807SJeff Garzik 5634c6fd2807SJeff Garzik poll_next = 0; 5635c6fd2807SJeff Garzik break; 5636c6fd2807SJeff Garzik default: 5637c6fd2807SJeff Garzik poll_next = 0; 5638c6fd2807SJeff Garzik BUG(); 5639c6fd2807SJeff Garzik } 5640c6fd2807SJeff Garzik 5641c6fd2807SJeff Garzik return poll_next; 5642c6fd2807SJeff Garzik } 5643c6fd2807SJeff Garzik 564465f27f38SDavid Howells static void ata_pio_task(struct work_struct *work) 5645c6fd2807SJeff Garzik { 564665f27f38SDavid Howells struct ata_port *ap = 564765f27f38SDavid Howells container_of(work, struct ata_port, port_task.work); 564865f27f38SDavid Howells struct ata_queued_cmd *qc = ap->port_task_data; 5649c6fd2807SJeff Garzik u8 status; 5650c6fd2807SJeff Garzik int poll_next; 5651c6fd2807SJeff Garzik 5652c6fd2807SJeff Garzik fsm_start: 5653c6fd2807SJeff Garzik WARN_ON(ap->hsm_task_state == HSM_ST_IDLE); 5654c6fd2807SJeff Garzik 5655c6fd2807SJeff Garzik /* 5656c6fd2807SJeff Garzik * This is purely heuristic. This is a fast path. 5657c6fd2807SJeff Garzik * Sometimes when we enter, BSY will be cleared in 5658c6fd2807SJeff Garzik * a chk-status or two. If not, the drive is probably seeking 5659c6fd2807SJeff Garzik * or something. Snooze for a couple msecs, then 5660c6fd2807SJeff Garzik * chk-status again. If still busy, queue delayed work. 5661c6fd2807SJeff Garzik */ 5662c6fd2807SJeff Garzik status = ata_busy_wait(ap, ATA_BUSY, 5); 5663c6fd2807SJeff Garzik if (status & ATA_BUSY) { 5664c6fd2807SJeff Garzik msleep(2); 5665c6fd2807SJeff Garzik status = ata_busy_wait(ap, ATA_BUSY, 10); 5666c6fd2807SJeff Garzik if (status & ATA_BUSY) { 5667442eacc3SJeff Garzik ata_pio_queue_task(ap, qc, ATA_SHORT_PAUSE); 5668c6fd2807SJeff Garzik return; 5669c6fd2807SJeff Garzik } 5670c6fd2807SJeff Garzik } 5671c6fd2807SJeff Garzik 5672c6fd2807SJeff Garzik /* move the HSM */ 5673c6fd2807SJeff Garzik poll_next = ata_hsm_move(ap, qc, status, 1); 5674c6fd2807SJeff Garzik 5675c6fd2807SJeff Garzik /* another command or interrupt handler 5676c6fd2807SJeff Garzik * may be running at this point. 5677c6fd2807SJeff Garzik */ 5678c6fd2807SJeff Garzik if (poll_next) 5679c6fd2807SJeff Garzik goto fsm_start; 5680c6fd2807SJeff Garzik } 5681c6fd2807SJeff Garzik 5682c6fd2807SJeff Garzik /** 5683c6fd2807SJeff Garzik * ata_qc_new - Request an available ATA command, for queueing 5684c6fd2807SJeff Garzik * @ap: Port associated with device @dev 5685c6fd2807SJeff Garzik * @dev: Device from whom we request an available command structure 5686c6fd2807SJeff Garzik * 5687c6fd2807SJeff Garzik * LOCKING: 5688c6fd2807SJeff Garzik * None. 5689c6fd2807SJeff Garzik */ 5690c6fd2807SJeff Garzik 5691c6fd2807SJeff Garzik static struct ata_queued_cmd *ata_qc_new(struct ata_port *ap) 5692c6fd2807SJeff Garzik { 5693c6fd2807SJeff Garzik struct ata_queued_cmd *qc = NULL; 5694c6fd2807SJeff Garzik unsigned int i; 5695c6fd2807SJeff Garzik 5696c6fd2807SJeff Garzik /* no command while frozen */ 5697c6fd2807SJeff Garzik if (unlikely(ap->pflags & ATA_PFLAG_FROZEN)) 5698c6fd2807SJeff Garzik return NULL; 5699c6fd2807SJeff Garzik 5700c6fd2807SJeff Garzik /* the last tag is reserved for internal command. */ 5701c6fd2807SJeff Garzik for (i = 0; i < ATA_MAX_QUEUE - 1; i++) 5702c6fd2807SJeff Garzik if (!test_and_set_bit(i, &ap->qc_allocated)) { 5703c6fd2807SJeff Garzik qc = __ata_qc_from_tag(ap, i); 5704c6fd2807SJeff Garzik break; 5705c6fd2807SJeff Garzik } 5706c6fd2807SJeff Garzik 5707c6fd2807SJeff Garzik if (qc) 5708c6fd2807SJeff Garzik qc->tag = i; 5709c6fd2807SJeff Garzik 5710c6fd2807SJeff Garzik return qc; 5711c6fd2807SJeff Garzik } 5712c6fd2807SJeff Garzik 5713c6fd2807SJeff Garzik /** 5714c6fd2807SJeff Garzik * ata_qc_new_init - Request an available ATA command, and initialize it 5715c6fd2807SJeff Garzik * @dev: Device from whom we request an available command structure 5716c6fd2807SJeff Garzik * 5717c6fd2807SJeff Garzik * LOCKING: 5718c6fd2807SJeff Garzik * None. 5719c6fd2807SJeff Garzik */ 5720c6fd2807SJeff Garzik 5721c6fd2807SJeff Garzik struct ata_queued_cmd *ata_qc_new_init(struct ata_device *dev) 5722c6fd2807SJeff Garzik { 57239af5c9c9STejun Heo struct ata_port *ap = dev->link->ap; 5724c6fd2807SJeff Garzik struct ata_queued_cmd *qc; 5725c6fd2807SJeff Garzik 5726c6fd2807SJeff Garzik qc = ata_qc_new(ap); 5727c6fd2807SJeff Garzik if (qc) { 5728c6fd2807SJeff Garzik qc->scsicmd = NULL; 5729c6fd2807SJeff Garzik qc->ap = ap; 5730c6fd2807SJeff Garzik qc->dev = dev; 5731c6fd2807SJeff Garzik 5732c6fd2807SJeff Garzik ata_qc_reinit(qc); 5733c6fd2807SJeff Garzik } 5734c6fd2807SJeff Garzik 5735c6fd2807SJeff Garzik return qc; 5736c6fd2807SJeff Garzik } 5737c6fd2807SJeff Garzik 5738c6fd2807SJeff Garzik /** 5739c6fd2807SJeff Garzik * ata_qc_free - free unused ata_queued_cmd 5740c6fd2807SJeff Garzik * @qc: Command to complete 5741c6fd2807SJeff Garzik * 5742c6fd2807SJeff Garzik * Designed to free unused ata_queued_cmd object 5743c6fd2807SJeff Garzik * in case something prevents using it. 5744c6fd2807SJeff Garzik * 5745c6fd2807SJeff Garzik * LOCKING: 5746cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 5747c6fd2807SJeff Garzik */ 5748c6fd2807SJeff Garzik void ata_qc_free(struct ata_queued_cmd *qc) 5749c6fd2807SJeff Garzik { 5750c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 5751c6fd2807SJeff Garzik unsigned int tag; 5752c6fd2807SJeff Garzik 5753c6fd2807SJeff Garzik WARN_ON(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ 5754c6fd2807SJeff Garzik 5755c6fd2807SJeff Garzik qc->flags = 0; 5756c6fd2807SJeff Garzik tag = qc->tag; 5757c6fd2807SJeff Garzik if (likely(ata_tag_valid(tag))) { 5758c6fd2807SJeff Garzik qc->tag = ATA_TAG_POISON; 5759c6fd2807SJeff Garzik clear_bit(tag, &ap->qc_allocated); 5760c6fd2807SJeff Garzik } 5761c6fd2807SJeff Garzik } 5762c6fd2807SJeff Garzik 5763c6fd2807SJeff Garzik void __ata_qc_complete(struct ata_queued_cmd *qc) 5764c6fd2807SJeff Garzik { 5765c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 57669af5c9c9STejun Heo struct ata_link *link = qc->dev->link; 5767c6fd2807SJeff Garzik 5768c6fd2807SJeff Garzik WARN_ON(qc == NULL); /* ata_qc_from_tag _might_ return NULL */ 5769c6fd2807SJeff Garzik WARN_ON(!(qc->flags & ATA_QCFLAG_ACTIVE)); 5770c6fd2807SJeff Garzik 5771c6fd2807SJeff Garzik if (likely(qc->flags & ATA_QCFLAG_DMAMAP)) 5772c6fd2807SJeff Garzik ata_sg_clean(qc); 5773c6fd2807SJeff Garzik 5774c6fd2807SJeff Garzik /* command should be marked inactive atomically with qc completion */ 5775da917d69STejun Heo if (qc->tf.protocol == ATA_PROT_NCQ) { 57769af5c9c9STejun Heo link->sactive &= ~(1 << qc->tag); 5777da917d69STejun Heo if (!link->sactive) 5778da917d69STejun Heo ap->nr_active_links--; 5779da917d69STejun Heo } else { 57809af5c9c9STejun Heo link->active_tag = ATA_TAG_POISON; 5781da917d69STejun Heo ap->nr_active_links--; 5782da917d69STejun Heo } 5783da917d69STejun Heo 5784da917d69STejun Heo /* clear exclusive status */ 5785da917d69STejun Heo if (unlikely(qc->flags & ATA_QCFLAG_CLEAR_EXCL && 5786da917d69STejun Heo ap->excl_link == link)) 5787da917d69STejun Heo ap->excl_link = NULL; 5788c6fd2807SJeff Garzik 5789c6fd2807SJeff Garzik /* atapi: mark qc as inactive to prevent the interrupt handler 5790c6fd2807SJeff Garzik * from completing the command twice later, before the error handler 5791c6fd2807SJeff Garzik * is called. (when rc != 0 and atapi request sense is needed) 5792c6fd2807SJeff Garzik */ 5793c6fd2807SJeff Garzik qc->flags &= ~ATA_QCFLAG_ACTIVE; 5794c6fd2807SJeff Garzik ap->qc_active &= ~(1 << qc->tag); 5795c6fd2807SJeff Garzik 5796c6fd2807SJeff Garzik /* call completion callback */ 5797c6fd2807SJeff Garzik qc->complete_fn(qc); 5798c6fd2807SJeff Garzik } 5799c6fd2807SJeff Garzik 580039599a53STejun Heo static void fill_result_tf(struct ata_queued_cmd *qc) 580139599a53STejun Heo { 580239599a53STejun Heo struct ata_port *ap = qc->ap; 580339599a53STejun Heo 580439599a53STejun Heo qc->result_tf.flags = qc->tf.flags; 58054742d54fSMark Lord ap->ops->tf_read(ap, &qc->result_tf); 580639599a53STejun Heo } 580739599a53STejun Heo 580800115e0fSTejun Heo static void ata_verify_xfer(struct ata_queued_cmd *qc) 580900115e0fSTejun Heo { 581000115e0fSTejun Heo struct ata_device *dev = qc->dev; 581100115e0fSTejun Heo 581200115e0fSTejun Heo if (ata_tag_internal(qc->tag)) 581300115e0fSTejun Heo return; 581400115e0fSTejun Heo 581500115e0fSTejun Heo if (ata_is_nodata(qc->tf.protocol)) 581600115e0fSTejun Heo return; 581700115e0fSTejun Heo 581800115e0fSTejun Heo if ((dev->mwdma_mask || dev->udma_mask) && ata_is_pio(qc->tf.protocol)) 581900115e0fSTejun Heo return; 582000115e0fSTejun Heo 582100115e0fSTejun Heo dev->flags &= ~ATA_DFLAG_DUBIOUS_XFER; 582200115e0fSTejun Heo } 582300115e0fSTejun Heo 5824c6fd2807SJeff Garzik /** 5825c6fd2807SJeff Garzik * ata_qc_complete - Complete an active ATA command 5826c6fd2807SJeff Garzik * @qc: Command to complete 5827c6fd2807SJeff Garzik * @err_mask: ATA Status register contents 5828c6fd2807SJeff Garzik * 5829c6fd2807SJeff Garzik * Indicate to the mid and upper layers that an ATA 5830c6fd2807SJeff Garzik * command has completed, with either an ok or not-ok status. 5831c6fd2807SJeff Garzik * 5832c6fd2807SJeff Garzik * LOCKING: 5833cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 5834c6fd2807SJeff Garzik */ 5835c6fd2807SJeff Garzik void ata_qc_complete(struct ata_queued_cmd *qc) 5836c6fd2807SJeff Garzik { 5837c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 5838c6fd2807SJeff Garzik 5839c6fd2807SJeff Garzik /* XXX: New EH and old EH use different mechanisms to 5840c6fd2807SJeff Garzik * synchronize EH with regular execution path. 5841c6fd2807SJeff Garzik * 5842c6fd2807SJeff Garzik * In new EH, a failed qc is marked with ATA_QCFLAG_FAILED. 5843c6fd2807SJeff Garzik * Normal execution path is responsible for not accessing a 5844c6fd2807SJeff Garzik * failed qc. libata core enforces the rule by returning NULL 5845c6fd2807SJeff Garzik * from ata_qc_from_tag() for failed qcs. 5846c6fd2807SJeff Garzik * 5847c6fd2807SJeff Garzik * Old EH depends on ata_qc_complete() nullifying completion 5848c6fd2807SJeff Garzik * requests if ATA_QCFLAG_EH_SCHEDULED is set. Old EH does 5849c6fd2807SJeff Garzik * not synchronize with interrupt handler. Only PIO task is 5850c6fd2807SJeff Garzik * taken care of. 5851c6fd2807SJeff Garzik */ 5852c6fd2807SJeff Garzik if (ap->ops->error_handler) { 58534dbfa39bSTejun Heo struct ata_device *dev = qc->dev; 58544dbfa39bSTejun Heo struct ata_eh_info *ehi = &dev->link->eh_info; 58554dbfa39bSTejun Heo 5856c6fd2807SJeff Garzik WARN_ON(ap->pflags & ATA_PFLAG_FROZEN); 5857c6fd2807SJeff Garzik 5858c6fd2807SJeff Garzik if (unlikely(qc->err_mask)) 5859c6fd2807SJeff Garzik qc->flags |= ATA_QCFLAG_FAILED; 5860c6fd2807SJeff Garzik 5861c6fd2807SJeff Garzik if (unlikely(qc->flags & ATA_QCFLAG_FAILED)) { 5862c6fd2807SJeff Garzik if (!ata_tag_internal(qc->tag)) { 5863c6fd2807SJeff Garzik /* always fill result TF for failed qc */ 586439599a53STejun Heo fill_result_tf(qc); 5865c6fd2807SJeff Garzik ata_qc_schedule_eh(qc); 5866c6fd2807SJeff Garzik return; 5867c6fd2807SJeff Garzik } 5868c6fd2807SJeff Garzik } 5869c6fd2807SJeff Garzik 5870c6fd2807SJeff Garzik /* read result TF if requested */ 5871c6fd2807SJeff Garzik if (qc->flags & ATA_QCFLAG_RESULT_TF) 587239599a53STejun Heo fill_result_tf(qc); 5873c6fd2807SJeff Garzik 58744dbfa39bSTejun Heo /* Some commands need post-processing after successful 58754dbfa39bSTejun Heo * completion. 58764dbfa39bSTejun Heo */ 58774dbfa39bSTejun Heo switch (qc->tf.command) { 58784dbfa39bSTejun Heo case ATA_CMD_SET_FEATURES: 58794dbfa39bSTejun Heo if (qc->tf.feature != SETFEATURES_WC_ON && 58804dbfa39bSTejun Heo qc->tf.feature != SETFEATURES_WC_OFF) 58814dbfa39bSTejun Heo break; 58824dbfa39bSTejun Heo /* fall through */ 58834dbfa39bSTejun Heo case ATA_CMD_INIT_DEV_PARAMS: /* CHS translation changed */ 58844dbfa39bSTejun Heo case ATA_CMD_SET_MULTI: /* multi_count changed */ 58854dbfa39bSTejun Heo /* revalidate device */ 58864dbfa39bSTejun Heo ehi->dev_action[dev->devno] |= ATA_EH_REVALIDATE; 58874dbfa39bSTejun Heo ata_port_schedule_eh(ap); 58884dbfa39bSTejun Heo break; 5889054a5fbaSTejun Heo 5890054a5fbaSTejun Heo case ATA_CMD_SLEEP: 5891054a5fbaSTejun Heo dev->flags |= ATA_DFLAG_SLEEPING; 5892054a5fbaSTejun Heo break; 58934dbfa39bSTejun Heo } 58944dbfa39bSTejun Heo 589500115e0fSTejun Heo if (unlikely(dev->flags & ATA_DFLAG_DUBIOUS_XFER)) 589600115e0fSTejun Heo ata_verify_xfer(qc); 589700115e0fSTejun Heo 5898c6fd2807SJeff Garzik __ata_qc_complete(qc); 5899c6fd2807SJeff Garzik } else { 5900c6fd2807SJeff Garzik if (qc->flags & ATA_QCFLAG_EH_SCHEDULED) 5901c6fd2807SJeff Garzik return; 5902c6fd2807SJeff Garzik 5903c6fd2807SJeff Garzik /* read result TF if failed or requested */ 5904c6fd2807SJeff Garzik if (qc->err_mask || qc->flags & ATA_QCFLAG_RESULT_TF) 590539599a53STejun Heo fill_result_tf(qc); 5906c6fd2807SJeff Garzik 5907c6fd2807SJeff Garzik __ata_qc_complete(qc); 5908c6fd2807SJeff Garzik } 5909c6fd2807SJeff Garzik } 5910c6fd2807SJeff Garzik 5911c6fd2807SJeff Garzik /** 5912c6fd2807SJeff Garzik * ata_qc_complete_multiple - Complete multiple qcs successfully 5913c6fd2807SJeff Garzik * @ap: port in question 5914c6fd2807SJeff Garzik * @qc_active: new qc_active mask 5915c6fd2807SJeff Garzik * @finish_qc: LLDD callback invoked before completing a qc 5916c6fd2807SJeff Garzik * 5917c6fd2807SJeff Garzik * Complete in-flight commands. This functions is meant to be 5918c6fd2807SJeff Garzik * called from low-level driver's interrupt routine to complete 5919c6fd2807SJeff Garzik * requests normally. ap->qc_active and @qc_active is compared 5920c6fd2807SJeff Garzik * and commands are completed accordingly. 5921c6fd2807SJeff Garzik * 5922c6fd2807SJeff Garzik * LOCKING: 5923cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 5924c6fd2807SJeff Garzik * 5925c6fd2807SJeff Garzik * RETURNS: 5926c6fd2807SJeff Garzik * Number of completed commands on success, -errno otherwise. 5927c6fd2807SJeff Garzik */ 5928c6fd2807SJeff Garzik int ata_qc_complete_multiple(struct ata_port *ap, u32 qc_active, 5929c6fd2807SJeff Garzik void (*finish_qc)(struct ata_queued_cmd *)) 5930c6fd2807SJeff Garzik { 5931c6fd2807SJeff Garzik int nr_done = 0; 5932c6fd2807SJeff Garzik u32 done_mask; 5933c6fd2807SJeff Garzik int i; 5934c6fd2807SJeff Garzik 5935c6fd2807SJeff Garzik done_mask = ap->qc_active ^ qc_active; 5936c6fd2807SJeff Garzik 5937c6fd2807SJeff Garzik if (unlikely(done_mask & qc_active)) { 5938c6fd2807SJeff Garzik ata_port_printk(ap, KERN_ERR, "illegal qc_active transition " 5939c6fd2807SJeff Garzik "(%08x->%08x)\n", ap->qc_active, qc_active); 5940c6fd2807SJeff Garzik return -EINVAL; 5941c6fd2807SJeff Garzik } 5942c6fd2807SJeff Garzik 5943c6fd2807SJeff Garzik for (i = 0; i < ATA_MAX_QUEUE; i++) { 5944c6fd2807SJeff Garzik struct ata_queued_cmd *qc; 5945c6fd2807SJeff Garzik 5946c6fd2807SJeff Garzik if (!(done_mask & (1 << i))) 5947c6fd2807SJeff Garzik continue; 5948c6fd2807SJeff Garzik 5949c6fd2807SJeff Garzik if ((qc = ata_qc_from_tag(ap, i))) { 5950c6fd2807SJeff Garzik if (finish_qc) 5951c6fd2807SJeff Garzik finish_qc(qc); 5952c6fd2807SJeff Garzik ata_qc_complete(qc); 5953c6fd2807SJeff Garzik nr_done++; 5954c6fd2807SJeff Garzik } 5955c6fd2807SJeff Garzik } 5956c6fd2807SJeff Garzik 5957c6fd2807SJeff Garzik return nr_done; 5958c6fd2807SJeff Garzik } 5959c6fd2807SJeff Garzik 5960c6fd2807SJeff Garzik /** 5961c6fd2807SJeff Garzik * ata_qc_issue - issue taskfile to device 5962c6fd2807SJeff Garzik * @qc: command to issue to device 5963c6fd2807SJeff Garzik * 5964c6fd2807SJeff Garzik * Prepare an ATA command to submission to device. 5965c6fd2807SJeff Garzik * This includes mapping the data into a DMA-able 5966c6fd2807SJeff Garzik * area, filling in the S/G table, and finally 5967c6fd2807SJeff Garzik * writing the taskfile to hardware, starting the command. 5968c6fd2807SJeff Garzik * 5969c6fd2807SJeff Garzik * LOCKING: 5970cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 5971c6fd2807SJeff Garzik */ 5972c6fd2807SJeff Garzik void ata_qc_issue(struct ata_queued_cmd *qc) 5973c6fd2807SJeff Garzik { 5974c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 59759af5c9c9STejun Heo struct ata_link *link = qc->dev->link; 5976405e66b3STejun Heo u8 prot = qc->tf.protocol; 5977c6fd2807SJeff Garzik 5978c6fd2807SJeff Garzik /* Make sure only one non-NCQ command is outstanding. The 5979c6fd2807SJeff Garzik * check is skipped for old EH because it reuses active qc to 5980c6fd2807SJeff Garzik * request ATAPI sense. 5981c6fd2807SJeff Garzik */ 59829af5c9c9STejun Heo WARN_ON(ap->ops->error_handler && ata_tag_valid(link->active_tag)); 5983c6fd2807SJeff Garzik 59841973a023STejun Heo if (ata_is_ncq(prot)) { 59859af5c9c9STejun Heo WARN_ON(link->sactive & (1 << qc->tag)); 5986da917d69STejun Heo 5987da917d69STejun Heo if (!link->sactive) 5988da917d69STejun Heo ap->nr_active_links++; 59899af5c9c9STejun Heo link->sactive |= 1 << qc->tag; 5990c6fd2807SJeff Garzik } else { 59919af5c9c9STejun Heo WARN_ON(link->sactive); 5992da917d69STejun Heo 5993da917d69STejun Heo ap->nr_active_links++; 59949af5c9c9STejun Heo link->active_tag = qc->tag; 5995c6fd2807SJeff Garzik } 5996c6fd2807SJeff Garzik 5997c6fd2807SJeff Garzik qc->flags |= ATA_QCFLAG_ACTIVE; 5998c6fd2807SJeff Garzik ap->qc_active |= 1 << qc->tag; 5999c6fd2807SJeff Garzik 6000f92a2636STejun Heo /* We guarantee to LLDs that they will have at least one 6001f92a2636STejun Heo * non-zero sg if the command is a data command. 6002f92a2636STejun Heo */ 6003ff2aeb1eSTejun Heo BUG_ON(ata_is_data(prot) && (!qc->sg || !qc->n_elem || !qc->nbytes)); 6004f92a2636STejun Heo 6005405e66b3STejun Heo if (ata_is_dma(prot) || (ata_is_pio(prot) && 6006f92a2636STejun Heo (ap->flags & ATA_FLAG_PIO_DMA))) 6007c6fd2807SJeff Garzik if (ata_sg_setup(qc)) 6008c6fd2807SJeff Garzik goto sg_err; 6009c6fd2807SJeff Garzik 6010054a5fbaSTejun Heo /* if device is sleeping, schedule softreset and abort the link */ 6011054a5fbaSTejun Heo if (unlikely(qc->dev->flags & ATA_DFLAG_SLEEPING)) { 6012054a5fbaSTejun Heo link->eh_info.action |= ATA_EH_SOFTRESET; 6013054a5fbaSTejun Heo ata_ehi_push_desc(&link->eh_info, "waking up from sleep"); 6014054a5fbaSTejun Heo ata_link_abort(link); 6015054a5fbaSTejun Heo return; 6016054a5fbaSTejun Heo } 6017054a5fbaSTejun Heo 6018c6fd2807SJeff Garzik ap->ops->qc_prep(qc); 6019c6fd2807SJeff Garzik 6020c6fd2807SJeff Garzik qc->err_mask |= ap->ops->qc_issue(qc); 6021c6fd2807SJeff Garzik if (unlikely(qc->err_mask)) 6022c6fd2807SJeff Garzik goto err; 6023c6fd2807SJeff Garzik return; 6024c6fd2807SJeff Garzik 6025c6fd2807SJeff Garzik sg_err: 6026c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_SYSTEM; 6027c6fd2807SJeff Garzik err: 6028c6fd2807SJeff Garzik ata_qc_complete(qc); 6029c6fd2807SJeff Garzik } 6030c6fd2807SJeff Garzik 6031c6fd2807SJeff Garzik /** 6032c6fd2807SJeff Garzik * ata_qc_issue_prot - issue taskfile to device in proto-dependent manner 6033c6fd2807SJeff Garzik * @qc: command to issue to device 6034c6fd2807SJeff Garzik * 6035c6fd2807SJeff Garzik * Using various libata functions and hooks, this function 6036c6fd2807SJeff Garzik * starts an ATA command. ATA commands are grouped into 6037c6fd2807SJeff Garzik * classes called "protocols", and issuing each type of protocol 6038c6fd2807SJeff Garzik * is slightly different. 6039c6fd2807SJeff Garzik * 6040c6fd2807SJeff Garzik * May be used as the qc_issue() entry in ata_port_operations. 6041c6fd2807SJeff Garzik * 6042c6fd2807SJeff Garzik * LOCKING: 6043cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 6044c6fd2807SJeff Garzik * 6045c6fd2807SJeff Garzik * RETURNS: 6046c6fd2807SJeff Garzik * Zero on success, AC_ERR_* mask on failure 6047c6fd2807SJeff Garzik */ 6048c6fd2807SJeff Garzik 6049c6fd2807SJeff Garzik unsigned int ata_qc_issue_prot(struct ata_queued_cmd *qc) 6050c6fd2807SJeff Garzik { 6051c6fd2807SJeff Garzik struct ata_port *ap = qc->ap; 6052c6fd2807SJeff Garzik 6053c6fd2807SJeff Garzik /* Use polling pio if the LLD doesn't handle 6054c6fd2807SJeff Garzik * interrupt driven pio and atapi CDB interrupt. 6055c6fd2807SJeff Garzik */ 6056c6fd2807SJeff Garzik if (ap->flags & ATA_FLAG_PIO_POLLING) { 6057c6fd2807SJeff Garzik switch (qc->tf.protocol) { 6058c6fd2807SJeff Garzik case ATA_PROT_PIO: 6059e3472cbeSAlbert Lee case ATA_PROT_NODATA: 60600dc36888STejun Heo case ATAPI_PROT_PIO: 60610dc36888STejun Heo case ATAPI_PROT_NODATA: 6062c6fd2807SJeff Garzik qc->tf.flags |= ATA_TFLAG_POLLING; 6063c6fd2807SJeff Garzik break; 60640dc36888STejun Heo case ATAPI_PROT_DMA: 6065c6fd2807SJeff Garzik if (qc->dev->flags & ATA_DFLAG_CDB_INTR) 6066c6fd2807SJeff Garzik /* see ata_dma_blacklisted() */ 6067c6fd2807SJeff Garzik BUG(); 6068c6fd2807SJeff Garzik break; 6069c6fd2807SJeff Garzik default: 6070c6fd2807SJeff Garzik break; 6071c6fd2807SJeff Garzik } 6072c6fd2807SJeff Garzik } 6073c6fd2807SJeff Garzik 6074c6fd2807SJeff Garzik /* select the device */ 6075c6fd2807SJeff Garzik ata_dev_select(ap, qc->dev->devno, 1, 0); 6076c6fd2807SJeff Garzik 6077c6fd2807SJeff Garzik /* start the command */ 6078c6fd2807SJeff Garzik switch (qc->tf.protocol) { 6079c6fd2807SJeff Garzik case ATA_PROT_NODATA: 6080c6fd2807SJeff Garzik if (qc->tf.flags & ATA_TFLAG_POLLING) 6081c6fd2807SJeff Garzik ata_qc_set_polling(qc); 6082c6fd2807SJeff Garzik 6083c6fd2807SJeff Garzik ata_tf_to_host(ap, &qc->tf); 6084c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_LAST; 6085c6fd2807SJeff Garzik 6086c6fd2807SJeff Garzik if (qc->tf.flags & ATA_TFLAG_POLLING) 6087442eacc3SJeff Garzik ata_pio_queue_task(ap, qc, 0); 6088c6fd2807SJeff Garzik 6089c6fd2807SJeff Garzik break; 6090c6fd2807SJeff Garzik 6091c6fd2807SJeff Garzik case ATA_PROT_DMA: 6092c6fd2807SJeff Garzik WARN_ON(qc->tf.flags & ATA_TFLAG_POLLING); 6093c6fd2807SJeff Garzik 6094c6fd2807SJeff Garzik ap->ops->tf_load(ap, &qc->tf); /* load tf registers */ 6095c6fd2807SJeff Garzik ap->ops->bmdma_setup(qc); /* set up bmdma */ 6096c6fd2807SJeff Garzik ap->ops->bmdma_start(qc); /* initiate bmdma */ 6097c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_LAST; 6098c6fd2807SJeff Garzik break; 6099c6fd2807SJeff Garzik 6100c6fd2807SJeff Garzik case ATA_PROT_PIO: 6101c6fd2807SJeff Garzik if (qc->tf.flags & ATA_TFLAG_POLLING) 6102c6fd2807SJeff Garzik ata_qc_set_polling(qc); 6103c6fd2807SJeff Garzik 6104c6fd2807SJeff Garzik ata_tf_to_host(ap, &qc->tf); 6105c6fd2807SJeff Garzik 6106c6fd2807SJeff Garzik if (qc->tf.flags & ATA_TFLAG_WRITE) { 6107c6fd2807SJeff Garzik /* PIO data out protocol */ 6108c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_FIRST; 6109442eacc3SJeff Garzik ata_pio_queue_task(ap, qc, 0); 6110c6fd2807SJeff Garzik 6111c6fd2807SJeff Garzik /* always send first data block using 6112c6fd2807SJeff Garzik * the ata_pio_task() codepath. 6113c6fd2807SJeff Garzik */ 6114c6fd2807SJeff Garzik } else { 6115c6fd2807SJeff Garzik /* PIO data in protocol */ 6116c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST; 6117c6fd2807SJeff Garzik 6118c6fd2807SJeff Garzik if (qc->tf.flags & ATA_TFLAG_POLLING) 6119442eacc3SJeff Garzik ata_pio_queue_task(ap, qc, 0); 6120c6fd2807SJeff Garzik 6121c6fd2807SJeff Garzik /* if polling, ata_pio_task() handles the rest. 6122c6fd2807SJeff Garzik * otherwise, interrupt handler takes over from here. 6123c6fd2807SJeff Garzik */ 6124c6fd2807SJeff Garzik } 6125c6fd2807SJeff Garzik 6126c6fd2807SJeff Garzik break; 6127c6fd2807SJeff Garzik 61280dc36888STejun Heo case ATAPI_PROT_PIO: 61290dc36888STejun Heo case ATAPI_PROT_NODATA: 6130c6fd2807SJeff Garzik if (qc->tf.flags & ATA_TFLAG_POLLING) 6131c6fd2807SJeff Garzik ata_qc_set_polling(qc); 6132c6fd2807SJeff Garzik 6133c6fd2807SJeff Garzik ata_tf_to_host(ap, &qc->tf); 6134c6fd2807SJeff Garzik 6135c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_FIRST; 6136c6fd2807SJeff Garzik 6137c6fd2807SJeff Garzik /* send cdb by polling if no cdb interrupt */ 6138c6fd2807SJeff Garzik if ((!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) || 6139c6fd2807SJeff Garzik (qc->tf.flags & ATA_TFLAG_POLLING)) 6140442eacc3SJeff Garzik ata_pio_queue_task(ap, qc, 0); 6141c6fd2807SJeff Garzik break; 6142c6fd2807SJeff Garzik 61430dc36888STejun Heo case ATAPI_PROT_DMA: 6144c6fd2807SJeff Garzik WARN_ON(qc->tf.flags & ATA_TFLAG_POLLING); 6145c6fd2807SJeff Garzik 6146c6fd2807SJeff Garzik ap->ops->tf_load(ap, &qc->tf); /* load tf registers */ 6147c6fd2807SJeff Garzik ap->ops->bmdma_setup(qc); /* set up bmdma */ 6148c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_FIRST; 6149c6fd2807SJeff Garzik 6150c6fd2807SJeff Garzik /* send cdb by polling if no cdb interrupt */ 6151c6fd2807SJeff Garzik if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) 6152442eacc3SJeff Garzik ata_pio_queue_task(ap, qc, 0); 6153c6fd2807SJeff Garzik break; 6154c6fd2807SJeff Garzik 6155c6fd2807SJeff Garzik default: 6156c6fd2807SJeff Garzik WARN_ON(1); 6157c6fd2807SJeff Garzik return AC_ERR_SYSTEM; 6158c6fd2807SJeff Garzik } 6159c6fd2807SJeff Garzik 6160c6fd2807SJeff Garzik return 0; 6161c6fd2807SJeff Garzik } 6162c6fd2807SJeff Garzik 6163c6fd2807SJeff Garzik /** 6164c6fd2807SJeff Garzik * ata_host_intr - Handle host interrupt for given (port, task) 6165c6fd2807SJeff Garzik * @ap: Port on which interrupt arrived (possibly...) 6166c6fd2807SJeff Garzik * @qc: Taskfile currently active in engine 6167c6fd2807SJeff Garzik * 6168c6fd2807SJeff Garzik * Handle host interrupt for given queued command. Currently, 6169c6fd2807SJeff Garzik * only DMA interrupts are handled. All other commands are 6170c6fd2807SJeff Garzik * handled via polling with interrupts disabled (nIEN bit). 6171c6fd2807SJeff Garzik * 6172c6fd2807SJeff Garzik * LOCKING: 6173cca3974eSJeff Garzik * spin_lock_irqsave(host lock) 6174c6fd2807SJeff Garzik * 6175c6fd2807SJeff Garzik * RETURNS: 6176c6fd2807SJeff Garzik * One if interrupt was handled, zero if not (shared irq). 6177c6fd2807SJeff Garzik */ 6178c6fd2807SJeff Garzik 6179c6fd2807SJeff Garzik inline unsigned int ata_host_intr(struct ata_port *ap, 6180c6fd2807SJeff Garzik struct ata_queued_cmd *qc) 6181c6fd2807SJeff Garzik { 61829af5c9c9STejun Heo struct ata_eh_info *ehi = &ap->link.eh_info; 6183c6fd2807SJeff Garzik u8 status, host_stat = 0; 6184c6fd2807SJeff Garzik 6185c6fd2807SJeff Garzik VPRINTK("ata%u: protocol %d task_state %d\n", 618644877b4eSTejun Heo ap->print_id, qc->tf.protocol, ap->hsm_task_state); 6187c6fd2807SJeff Garzik 6188c6fd2807SJeff Garzik /* Check whether we are expecting interrupt in this state */ 6189c6fd2807SJeff Garzik switch (ap->hsm_task_state) { 6190c6fd2807SJeff Garzik case HSM_ST_FIRST: 6191c6fd2807SJeff Garzik /* Some pre-ATAPI-4 devices assert INTRQ 6192c6fd2807SJeff Garzik * at this state when ready to receive CDB. 6193c6fd2807SJeff Garzik */ 6194c6fd2807SJeff Garzik 6195c6fd2807SJeff Garzik /* Check the ATA_DFLAG_CDB_INTR flag is enough here. 6196405e66b3STejun Heo * The flag was turned on only for atapi devices. No 6197405e66b3STejun Heo * need to check ata_is_atapi(qc->tf.protocol) again. 6198c6fd2807SJeff Garzik */ 6199c6fd2807SJeff Garzik if (!(qc->dev->flags & ATA_DFLAG_CDB_INTR)) 6200c6fd2807SJeff Garzik goto idle_irq; 6201c6fd2807SJeff Garzik break; 6202c6fd2807SJeff Garzik case HSM_ST_LAST: 6203c6fd2807SJeff Garzik if (qc->tf.protocol == ATA_PROT_DMA || 62040dc36888STejun Heo qc->tf.protocol == ATAPI_PROT_DMA) { 6205c6fd2807SJeff Garzik /* check status of DMA engine */ 6206c6fd2807SJeff Garzik host_stat = ap->ops->bmdma_status(ap); 620744877b4eSTejun Heo VPRINTK("ata%u: host_stat 0x%X\n", 620844877b4eSTejun Heo ap->print_id, host_stat); 6209c6fd2807SJeff Garzik 6210c6fd2807SJeff Garzik /* if it's not our irq... */ 6211c6fd2807SJeff Garzik if (!(host_stat & ATA_DMA_INTR)) 6212c6fd2807SJeff Garzik goto idle_irq; 6213c6fd2807SJeff Garzik 6214c6fd2807SJeff Garzik /* before we do anything else, clear DMA-Start bit */ 6215c6fd2807SJeff Garzik ap->ops->bmdma_stop(qc); 6216c6fd2807SJeff Garzik 6217c6fd2807SJeff Garzik if (unlikely(host_stat & ATA_DMA_ERR)) { 6218c6fd2807SJeff Garzik /* error when transfering data to/from memory */ 6219c6fd2807SJeff Garzik qc->err_mask |= AC_ERR_HOST_BUS; 6220c6fd2807SJeff Garzik ap->hsm_task_state = HSM_ST_ERR; 6221c6fd2807SJeff Garzik } 6222c6fd2807SJeff Garzik } 6223c6fd2807SJeff Garzik break; 6224c6fd2807SJeff Garzik case HSM_ST: 6225c6fd2807SJeff Garzik break; 6226c6fd2807SJeff Garzik default: 6227c6fd2807SJeff Garzik goto idle_irq; 6228c6fd2807SJeff Garzik } 6229c6fd2807SJeff Garzik 6230c6fd2807SJeff Garzik /* check altstatus */ 6231c6fd2807SJeff Garzik status = ata_altstatus(ap); 6232c6fd2807SJeff Garzik if (status & ATA_BUSY) 6233c6fd2807SJeff Garzik goto idle_irq; 6234c6fd2807SJeff Garzik 6235c6fd2807SJeff Garzik /* check main status, clearing INTRQ */ 6236c6fd2807SJeff Garzik status = ata_chk_status(ap); 6237c6fd2807SJeff Garzik if (unlikely(status & ATA_BUSY)) 6238c6fd2807SJeff Garzik goto idle_irq; 6239c6fd2807SJeff Garzik 6240c6fd2807SJeff Garzik /* ack bmdma irq events */ 6241c6fd2807SJeff Garzik ap->ops->irq_clear(ap); 6242c6fd2807SJeff Garzik 6243c6fd2807SJeff Garzik ata_hsm_move(ap, qc, status, 0); 6244ea54763fSTejun Heo 6245ea54763fSTejun Heo if (unlikely(qc->err_mask) && (qc->tf.protocol == ATA_PROT_DMA || 62460dc36888STejun Heo qc->tf.protocol == ATAPI_PROT_DMA)) 6247ea54763fSTejun Heo ata_ehi_push_desc(ehi, "BMDMA stat 0x%x", host_stat); 6248ea54763fSTejun Heo 6249c6fd2807SJeff Garzik return 1; /* irq handled */ 6250c6fd2807SJeff Garzik 6251c6fd2807SJeff Garzik idle_irq: 6252c6fd2807SJeff Garzik ap->stats.idle_irq++; 6253c6fd2807SJeff Garzik 6254c6fd2807SJeff Garzik #ifdef ATA_IRQ_TRAP 6255c6fd2807SJeff Garzik if ((ap->stats.idle_irq % 1000) == 0) { 62566d32d30fSJeff Garzik ata_chk_status(ap); 62576d32d30fSJeff Garzik ap->ops->irq_clear(ap); 6258c6fd2807SJeff Garzik ata_port_printk(ap, KERN_WARNING, "irq trap\n"); 6259c6fd2807SJeff Garzik return 1; 6260c6fd2807SJeff Garzik } 6261c6fd2807SJeff Garzik #endif 6262c6fd2807SJeff Garzik return 0; /* irq not handled */ 6263c6fd2807SJeff Garzik } 6264c6fd2807SJeff Garzik 6265c6fd2807SJeff Garzik /** 6266c6fd2807SJeff Garzik * ata_interrupt - Default ATA host interrupt handler 6267c6fd2807SJeff Garzik * @irq: irq line (unused) 6268cca3974eSJeff Garzik * @dev_instance: pointer to our ata_host information structure 6269c6fd2807SJeff Garzik * 6270c6fd2807SJeff Garzik * Default interrupt handler for PCI IDE devices. Calls 6271c6fd2807SJeff Garzik * ata_host_intr() for each port that is not disabled. 6272c6fd2807SJeff Garzik * 6273c6fd2807SJeff Garzik * LOCKING: 6274cca3974eSJeff Garzik * Obtains host lock during operation. 6275c6fd2807SJeff Garzik * 6276c6fd2807SJeff Garzik * RETURNS: 6277c6fd2807SJeff Garzik * IRQ_NONE or IRQ_HANDLED. 6278c6fd2807SJeff Garzik */ 6279c6fd2807SJeff Garzik 62807d12e780SDavid Howells irqreturn_t ata_interrupt(int irq, void *dev_instance) 6281c6fd2807SJeff Garzik { 6282cca3974eSJeff Garzik struct ata_host *host = dev_instance; 6283c6fd2807SJeff Garzik unsigned int i; 6284c6fd2807SJeff Garzik unsigned int handled = 0; 6285c6fd2807SJeff Garzik unsigned long flags; 6286c6fd2807SJeff Garzik 6287c6fd2807SJeff Garzik /* TODO: make _irqsave conditional on x86 PCI IDE legacy mode */ 6288cca3974eSJeff Garzik spin_lock_irqsave(&host->lock, flags); 6289c6fd2807SJeff Garzik 6290cca3974eSJeff Garzik for (i = 0; i < host->n_ports; i++) { 6291c6fd2807SJeff Garzik struct ata_port *ap; 6292c6fd2807SJeff Garzik 6293cca3974eSJeff Garzik ap = host->ports[i]; 6294c6fd2807SJeff Garzik if (ap && 6295c6fd2807SJeff Garzik !(ap->flags & ATA_FLAG_DISABLED)) { 6296c6fd2807SJeff Garzik struct ata_queued_cmd *qc; 6297c6fd2807SJeff Garzik 62989af5c9c9STejun Heo qc = ata_qc_from_tag(ap, ap->link.active_tag); 6299c6fd2807SJeff Garzik if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING)) && 6300c6fd2807SJeff Garzik (qc->flags & ATA_QCFLAG_ACTIVE)) 6301c6fd2807SJeff Garzik handled |= ata_host_intr(ap, qc); 6302c6fd2807SJeff Garzik } 6303c6fd2807SJeff Garzik } 6304c6fd2807SJeff Garzik 6305cca3974eSJeff Garzik spin_unlock_irqrestore(&host->lock, flags); 6306c6fd2807SJeff Garzik 6307c6fd2807SJeff Garzik return IRQ_RETVAL(handled); 6308c6fd2807SJeff Garzik } 6309c6fd2807SJeff Garzik 6310c6fd2807SJeff Garzik /** 6311c6fd2807SJeff Garzik * sata_scr_valid - test whether SCRs are accessible 6312936fd732STejun Heo * @link: ATA link to test SCR accessibility for 6313c6fd2807SJeff Garzik * 6314936fd732STejun Heo * Test whether SCRs are accessible for @link. 6315c6fd2807SJeff Garzik * 6316c6fd2807SJeff Garzik * LOCKING: 6317c6fd2807SJeff Garzik * None. 6318c6fd2807SJeff Garzik * 6319c6fd2807SJeff Garzik * RETURNS: 6320c6fd2807SJeff Garzik * 1 if SCRs are accessible, 0 otherwise. 6321c6fd2807SJeff Garzik */ 6322936fd732STejun Heo int sata_scr_valid(struct ata_link *link) 6323c6fd2807SJeff Garzik { 6324936fd732STejun Heo struct ata_port *ap = link->ap; 6325936fd732STejun Heo 6326a16abc0bSTejun Heo return (ap->flags & ATA_FLAG_SATA) && ap->ops->scr_read; 6327c6fd2807SJeff Garzik } 6328c6fd2807SJeff Garzik 6329c6fd2807SJeff Garzik /** 6330c6fd2807SJeff Garzik * sata_scr_read - read SCR register of the specified port 6331936fd732STejun Heo * @link: ATA link to read SCR for 6332c6fd2807SJeff Garzik * @reg: SCR to read 6333c6fd2807SJeff Garzik * @val: Place to store read value 6334c6fd2807SJeff Garzik * 6335936fd732STejun Heo * Read SCR register @reg of @link into *@val. This function is 6336633273a3STejun Heo * guaranteed to succeed if @link is ap->link, the cable type of 6337633273a3STejun Heo * the port is SATA and the port implements ->scr_read. 6338c6fd2807SJeff Garzik * 6339c6fd2807SJeff Garzik * LOCKING: 6340633273a3STejun Heo * None if @link is ap->link. Kernel thread context otherwise. 6341c6fd2807SJeff Garzik * 6342c6fd2807SJeff Garzik * RETURNS: 6343c6fd2807SJeff Garzik * 0 on success, negative errno on failure. 6344c6fd2807SJeff Garzik */ 6345936fd732STejun Heo int sata_scr_read(struct ata_link *link, int reg, u32 *val) 6346c6fd2807SJeff Garzik { 6347633273a3STejun Heo if (ata_is_host_link(link)) { 6348936fd732STejun Heo struct ata_port *ap = link->ap; 6349936fd732STejun Heo 6350936fd732STejun Heo if (sata_scr_valid(link)) 6351da3dbb17STejun Heo return ap->ops->scr_read(ap, reg, val); 6352c6fd2807SJeff Garzik return -EOPNOTSUPP; 6353c6fd2807SJeff Garzik } 6354c6fd2807SJeff Garzik 6355633273a3STejun Heo return sata_pmp_scr_read(link, reg, val); 6356633273a3STejun Heo } 6357633273a3STejun Heo 6358c6fd2807SJeff Garzik /** 6359c6fd2807SJeff Garzik * sata_scr_write - write SCR register of the specified port 6360936fd732STejun Heo * @link: ATA link to write SCR for 6361c6fd2807SJeff Garzik * @reg: SCR to write 6362c6fd2807SJeff Garzik * @val: value to write 6363c6fd2807SJeff Garzik * 6364936fd732STejun Heo * Write @val to SCR register @reg of @link. This function is 6365633273a3STejun Heo * guaranteed to succeed if @link is ap->link, the cable type of 6366633273a3STejun Heo * the port is SATA and the port implements ->scr_read. 6367c6fd2807SJeff Garzik * 6368c6fd2807SJeff Garzik * LOCKING: 6369633273a3STejun Heo * None if @link is ap->link. Kernel thread context otherwise. 6370c6fd2807SJeff Garzik * 6371c6fd2807SJeff Garzik * RETURNS: 6372c6fd2807SJeff Garzik * 0 on success, negative errno on failure. 6373c6fd2807SJeff Garzik */ 6374936fd732STejun Heo int sata_scr_write(struct ata_link *link, int reg, u32 val) 6375c6fd2807SJeff Garzik { 6376633273a3STejun Heo if (ata_is_host_link(link)) { 6377936fd732STejun Heo struct ata_port *ap = link->ap; 6378936fd732STejun Heo 6379936fd732STejun Heo if (sata_scr_valid(link)) 6380da3dbb17STejun Heo return ap->ops->scr_write(ap, reg, val); 6381c6fd2807SJeff Garzik return -EOPNOTSUPP; 6382c6fd2807SJeff Garzik } 6383c6fd2807SJeff Garzik 6384633273a3STejun Heo return sata_pmp_scr_write(link, reg, val); 6385633273a3STejun Heo } 6386633273a3STejun Heo 6387c6fd2807SJeff Garzik /** 6388c6fd2807SJeff Garzik * sata_scr_write_flush - write SCR register of the specified port and flush 6389936fd732STejun Heo * @link: ATA link to write SCR for 6390c6fd2807SJeff Garzik * @reg: SCR to write 6391c6fd2807SJeff Garzik * @val: value to write 6392c6fd2807SJeff Garzik * 6393c6fd2807SJeff Garzik * This function is identical to sata_scr_write() except that this 6394c6fd2807SJeff Garzik * function performs flush after writing to the register. 6395c6fd2807SJeff Garzik * 6396c6fd2807SJeff Garzik * LOCKING: 6397633273a3STejun Heo * None if @link is ap->link. Kernel thread context otherwise. 6398c6fd2807SJeff Garzik * 6399c6fd2807SJeff Garzik * RETURNS: 6400c6fd2807SJeff Garzik * 0 on success, negative errno on failure. 6401c6fd2807SJeff Garzik */ 6402936fd732STejun Heo int sata_scr_write_flush(struct ata_link *link, int reg, u32 val) 6403c6fd2807SJeff Garzik { 6404633273a3STejun Heo if (ata_is_host_link(link)) { 6405936fd732STejun Heo struct ata_port *ap = link->ap; 6406da3dbb17STejun Heo int rc; 6407da3dbb17STejun Heo 6408936fd732STejun Heo if (sata_scr_valid(link)) { 6409da3dbb17STejun Heo rc = ap->ops->scr_write(ap, reg, val); 6410da3dbb17STejun Heo if (rc == 0) 6411da3dbb17STejun Heo rc = ap->ops->scr_read(ap, reg, &val); 6412da3dbb17STejun Heo return rc; 6413c6fd2807SJeff Garzik } 6414c6fd2807SJeff Garzik return -EOPNOTSUPP; 6415c6fd2807SJeff Garzik } 6416c6fd2807SJeff Garzik 6417633273a3STejun Heo return sata_pmp_scr_write(link, reg, val); 6418633273a3STejun Heo } 6419633273a3STejun Heo 6420c6fd2807SJeff Garzik /** 6421936fd732STejun Heo * ata_link_online - test whether the given link is online 6422936fd732STejun Heo * @link: ATA link to test 6423c6fd2807SJeff Garzik * 6424936fd732STejun Heo * Test whether @link is online. Note that this function returns 6425936fd732STejun Heo * 0 if online status of @link cannot be obtained, so 6426936fd732STejun Heo * ata_link_online(link) != !ata_link_offline(link). 6427c6fd2807SJeff Garzik * 6428c6fd2807SJeff Garzik * LOCKING: 6429c6fd2807SJeff Garzik * None. 6430c6fd2807SJeff Garzik * 6431c6fd2807SJeff Garzik * RETURNS: 6432c6fd2807SJeff Garzik * 1 if the port online status is available and online. 6433c6fd2807SJeff Garzik */ 6434936fd732STejun Heo int ata_link_online(struct ata_link *link) 6435c6fd2807SJeff Garzik { 6436c6fd2807SJeff Garzik u32 sstatus; 6437c6fd2807SJeff Garzik 6438936fd732STejun Heo if (sata_scr_read(link, SCR_STATUS, &sstatus) == 0 && 6439936fd732STejun Heo (sstatus & 0xf) == 0x3) 6440c6fd2807SJeff Garzik return 1; 6441c6fd2807SJeff Garzik return 0; 6442c6fd2807SJeff Garzik } 6443c6fd2807SJeff Garzik 6444c6fd2807SJeff Garzik /** 6445936fd732STejun Heo * ata_link_offline - test whether the given link is offline 6446936fd732STejun Heo * @link: ATA link to test 6447c6fd2807SJeff Garzik * 6448936fd732STejun Heo * Test whether @link is offline. Note that this function 6449936fd732STejun Heo * returns 0 if offline status of @link cannot be obtained, so 6450936fd732STejun Heo * ata_link_online(link) != !ata_link_offline(link). 6451c6fd2807SJeff Garzik * 6452c6fd2807SJeff Garzik * LOCKING: 6453c6fd2807SJeff Garzik * None. 6454c6fd2807SJeff Garzik * 6455c6fd2807SJeff Garzik * RETURNS: 6456c6fd2807SJeff Garzik * 1 if the port offline status is available and offline. 6457c6fd2807SJeff Garzik */ 6458936fd732STejun Heo int ata_link_offline(struct ata_link *link) 6459c6fd2807SJeff Garzik { 6460c6fd2807SJeff Garzik u32 sstatus; 6461c6fd2807SJeff Garzik 6462936fd732STejun Heo if (sata_scr_read(link, SCR_STATUS, &sstatus) == 0 && 6463936fd732STejun Heo (sstatus & 0xf) != 0x3) 6464c6fd2807SJeff Garzik return 1; 6465c6fd2807SJeff Garzik return 0; 6466c6fd2807SJeff Garzik } 6467c6fd2807SJeff Garzik 6468c6fd2807SJeff Garzik int ata_flush_cache(struct ata_device *dev) 6469c6fd2807SJeff Garzik { 6470c6fd2807SJeff Garzik unsigned int err_mask; 6471c6fd2807SJeff Garzik u8 cmd; 6472c6fd2807SJeff Garzik 6473c6fd2807SJeff Garzik if (!ata_try_flush_cache(dev)) 6474c6fd2807SJeff Garzik return 0; 6475c6fd2807SJeff Garzik 64766fc49adbSTejun Heo if (dev->flags & ATA_DFLAG_FLUSH_EXT) 6477c6fd2807SJeff Garzik cmd = ATA_CMD_FLUSH_EXT; 6478c6fd2807SJeff Garzik else 6479c6fd2807SJeff Garzik cmd = ATA_CMD_FLUSH; 6480c6fd2807SJeff Garzik 64814f34337bSAlan Cox /* This is wrong. On a failed flush we get back the LBA of the lost 64824f34337bSAlan Cox sector and we should (assuming it wasn't aborted as unknown) issue 64834f34337bSAlan Cox a further flush command to continue the writeback until it 64844f34337bSAlan Cox does not error */ 6485c6fd2807SJeff Garzik err_mask = ata_do_simple_cmd(dev, cmd); 6486c6fd2807SJeff Garzik if (err_mask) { 6487c6fd2807SJeff Garzik ata_dev_printk(dev, KERN_ERR, "failed to flush cache\n"); 6488c6fd2807SJeff Garzik return -EIO; 6489c6fd2807SJeff Garzik } 6490c6fd2807SJeff Garzik 6491c6fd2807SJeff Garzik return 0; 6492c6fd2807SJeff Garzik } 6493c6fd2807SJeff Garzik 64946ffa01d8STejun Heo #ifdef CONFIG_PM 6495cca3974eSJeff Garzik static int ata_host_request_pm(struct ata_host *host, pm_message_t mesg, 6496cca3974eSJeff Garzik unsigned int action, unsigned int ehi_flags, 6497cca3974eSJeff Garzik int wait) 6498c6fd2807SJeff Garzik { 6499c6fd2807SJeff Garzik unsigned long flags; 6500c6fd2807SJeff Garzik int i, rc; 6501c6fd2807SJeff Garzik 6502cca3974eSJeff Garzik for (i = 0; i < host->n_ports; i++) { 6503cca3974eSJeff Garzik struct ata_port *ap = host->ports[i]; 6504e3667ebfSTejun Heo struct ata_link *link; 6505c6fd2807SJeff Garzik 6506c6fd2807SJeff Garzik /* Previous resume operation might still be in 6507c6fd2807SJeff Garzik * progress. Wait for PM_PENDING to clear. 6508c6fd2807SJeff Garzik */ 6509c6fd2807SJeff Garzik if (ap->pflags & ATA_PFLAG_PM_PENDING) { 6510c6fd2807SJeff Garzik ata_port_wait_eh(ap); 6511c6fd2807SJeff Garzik WARN_ON(ap->pflags & ATA_PFLAG_PM_PENDING); 6512c6fd2807SJeff Garzik } 6513c6fd2807SJeff Garzik 6514c6fd2807SJeff Garzik /* request PM ops to EH */ 6515c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 6516c6fd2807SJeff Garzik 6517c6fd2807SJeff Garzik ap->pm_mesg = mesg; 6518c6fd2807SJeff Garzik if (wait) { 6519c6fd2807SJeff Garzik rc = 0; 6520c6fd2807SJeff Garzik ap->pm_result = &rc; 6521c6fd2807SJeff Garzik } 6522c6fd2807SJeff Garzik 6523c6fd2807SJeff Garzik ap->pflags |= ATA_PFLAG_PM_PENDING; 6524e3667ebfSTejun Heo __ata_port_for_each_link(link, ap) { 6525e3667ebfSTejun Heo link->eh_info.action |= action; 6526e3667ebfSTejun Heo link->eh_info.flags |= ehi_flags; 6527e3667ebfSTejun Heo } 6528c6fd2807SJeff Garzik 6529c6fd2807SJeff Garzik ata_port_schedule_eh(ap); 6530c6fd2807SJeff Garzik 6531c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 6532c6fd2807SJeff Garzik 6533c6fd2807SJeff Garzik /* wait and check result */ 6534c6fd2807SJeff Garzik if (wait) { 6535c6fd2807SJeff Garzik ata_port_wait_eh(ap); 6536c6fd2807SJeff Garzik WARN_ON(ap->pflags & ATA_PFLAG_PM_PENDING); 6537c6fd2807SJeff Garzik if (rc) 6538c6fd2807SJeff Garzik return rc; 6539c6fd2807SJeff Garzik } 6540c6fd2807SJeff Garzik } 6541c6fd2807SJeff Garzik 6542c6fd2807SJeff Garzik return 0; 6543c6fd2807SJeff Garzik } 6544c6fd2807SJeff Garzik 6545c6fd2807SJeff Garzik /** 6546cca3974eSJeff Garzik * ata_host_suspend - suspend host 6547cca3974eSJeff Garzik * @host: host to suspend 6548c6fd2807SJeff Garzik * @mesg: PM message 6549c6fd2807SJeff Garzik * 6550cca3974eSJeff Garzik * Suspend @host. Actual operation is performed by EH. This 6551c6fd2807SJeff Garzik * function requests EH to perform PM operations and waits for EH 6552c6fd2807SJeff Garzik * to finish. 6553c6fd2807SJeff Garzik * 6554c6fd2807SJeff Garzik * LOCKING: 6555c6fd2807SJeff Garzik * Kernel thread context (may sleep). 6556c6fd2807SJeff Garzik * 6557c6fd2807SJeff Garzik * RETURNS: 6558c6fd2807SJeff Garzik * 0 on success, -errno on failure. 6559c6fd2807SJeff Garzik */ 6560cca3974eSJeff Garzik int ata_host_suspend(struct ata_host *host, pm_message_t mesg) 6561c6fd2807SJeff Garzik { 65629666f400STejun Heo int rc; 6563c6fd2807SJeff Garzik 6564ca77329fSKristen Carlson Accardi /* 6565ca77329fSKristen Carlson Accardi * disable link pm on all ports before requesting 6566ca77329fSKristen Carlson Accardi * any pm activity 6567ca77329fSKristen Carlson Accardi */ 6568ca77329fSKristen Carlson Accardi ata_lpm_enable(host); 6569ca77329fSKristen Carlson Accardi 6570cca3974eSJeff Garzik rc = ata_host_request_pm(host, mesg, 0, ATA_EHI_QUIET, 1); 657172ad6ec4SJeff Garzik if (rc == 0) 657272ad6ec4SJeff Garzik host->dev->power.power_state = mesg; 6573c6fd2807SJeff Garzik return rc; 6574c6fd2807SJeff Garzik } 6575c6fd2807SJeff Garzik 6576c6fd2807SJeff Garzik /** 6577cca3974eSJeff Garzik * ata_host_resume - resume host 6578cca3974eSJeff Garzik * @host: host to resume 6579c6fd2807SJeff Garzik * 6580cca3974eSJeff Garzik * Resume @host. Actual operation is performed by EH. This 6581c6fd2807SJeff Garzik * function requests EH to perform PM operations and returns. 6582c6fd2807SJeff Garzik * Note that all resume operations are performed parallely. 6583c6fd2807SJeff Garzik * 6584c6fd2807SJeff Garzik * LOCKING: 6585c6fd2807SJeff Garzik * Kernel thread context (may sleep). 6586c6fd2807SJeff Garzik */ 6587cca3974eSJeff Garzik void ata_host_resume(struct ata_host *host) 6588c6fd2807SJeff Garzik { 6589cca3974eSJeff Garzik ata_host_request_pm(host, PMSG_ON, ATA_EH_SOFTRESET, 6590c6fd2807SJeff Garzik ATA_EHI_NO_AUTOPSY | ATA_EHI_QUIET, 0); 659172ad6ec4SJeff Garzik host->dev->power.power_state = PMSG_ON; 6592ca77329fSKristen Carlson Accardi 6593ca77329fSKristen Carlson Accardi /* reenable link pm */ 6594ca77329fSKristen Carlson Accardi ata_lpm_disable(host); 6595c6fd2807SJeff Garzik } 65966ffa01d8STejun Heo #endif 6597c6fd2807SJeff Garzik 6598c6fd2807SJeff Garzik /** 6599c6fd2807SJeff Garzik * ata_port_start - Set port up for dma. 6600c6fd2807SJeff Garzik * @ap: Port to initialize 6601c6fd2807SJeff Garzik * 6602c6fd2807SJeff Garzik * Called just after data structures for each port are 6603c6fd2807SJeff Garzik * initialized. Allocates space for PRD table. 6604c6fd2807SJeff Garzik * 6605c6fd2807SJeff Garzik * May be used as the port_start() entry in ata_port_operations. 6606c6fd2807SJeff Garzik * 6607c6fd2807SJeff Garzik * LOCKING: 6608c6fd2807SJeff Garzik * Inherited from caller. 6609c6fd2807SJeff Garzik */ 6610c6fd2807SJeff Garzik int ata_port_start(struct ata_port *ap) 6611c6fd2807SJeff Garzik { 6612c6fd2807SJeff Garzik struct device *dev = ap->dev; 6613c6fd2807SJeff Garzik 6614f0d36efdSTejun Heo ap->prd = dmam_alloc_coherent(dev, ATA_PRD_TBL_SZ, &ap->prd_dma, 6615f0d36efdSTejun Heo GFP_KERNEL); 6616c6fd2807SJeff Garzik if (!ap->prd) 6617c6fd2807SJeff Garzik return -ENOMEM; 6618c6fd2807SJeff Garzik 6619c6fd2807SJeff Garzik return 0; 6620c6fd2807SJeff Garzik } 6621c6fd2807SJeff Garzik 6622c6fd2807SJeff Garzik /** 6623c6fd2807SJeff Garzik * ata_dev_init - Initialize an ata_device structure 6624c6fd2807SJeff Garzik * @dev: Device structure to initialize 6625c6fd2807SJeff Garzik * 6626c6fd2807SJeff Garzik * Initialize @dev in preparation for probing. 6627c6fd2807SJeff Garzik * 6628c6fd2807SJeff Garzik * LOCKING: 6629c6fd2807SJeff Garzik * Inherited from caller. 6630c6fd2807SJeff Garzik */ 6631c6fd2807SJeff Garzik void ata_dev_init(struct ata_device *dev) 6632c6fd2807SJeff Garzik { 66339af5c9c9STejun Heo struct ata_link *link = dev->link; 66349af5c9c9STejun Heo struct ata_port *ap = link->ap; 6635c6fd2807SJeff Garzik unsigned long flags; 6636c6fd2807SJeff Garzik 6637c6fd2807SJeff Garzik /* SATA spd limit is bound to the first device */ 66389af5c9c9STejun Heo link->sata_spd_limit = link->hw_sata_spd_limit; 66399af5c9c9STejun Heo link->sata_spd = 0; 6640c6fd2807SJeff Garzik 6641c6fd2807SJeff Garzik /* High bits of dev->flags are used to record warm plug 6642c6fd2807SJeff Garzik * requests which occur asynchronously. Synchronize using 6643cca3974eSJeff Garzik * host lock. 6644c6fd2807SJeff Garzik */ 6645c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 6646c6fd2807SJeff Garzik dev->flags &= ~ATA_DFLAG_INIT_MASK; 66473dcc323fSTejun Heo dev->horkage = 0; 6648c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 6649c6fd2807SJeff Garzik 6650c6fd2807SJeff Garzik memset((void *)dev + ATA_DEVICE_CLEAR_OFFSET, 0, 6651c6fd2807SJeff Garzik sizeof(*dev) - ATA_DEVICE_CLEAR_OFFSET); 6652c6fd2807SJeff Garzik dev->pio_mask = UINT_MAX; 6653c6fd2807SJeff Garzik dev->mwdma_mask = UINT_MAX; 6654c6fd2807SJeff Garzik dev->udma_mask = UINT_MAX; 6655c6fd2807SJeff Garzik } 6656c6fd2807SJeff Garzik 6657c6fd2807SJeff Garzik /** 66584fb37a25STejun Heo * ata_link_init - Initialize an ata_link structure 66594fb37a25STejun Heo * @ap: ATA port link is attached to 66604fb37a25STejun Heo * @link: Link structure to initialize 66618989805dSTejun Heo * @pmp: Port multiplier port number 66624fb37a25STejun Heo * 66634fb37a25STejun Heo * Initialize @link. 66644fb37a25STejun Heo * 66654fb37a25STejun Heo * LOCKING: 66664fb37a25STejun Heo * Kernel thread context (may sleep) 66674fb37a25STejun Heo */ 6668fb7fd614STejun Heo void ata_link_init(struct ata_port *ap, struct ata_link *link, int pmp) 66694fb37a25STejun Heo { 66704fb37a25STejun Heo int i; 66714fb37a25STejun Heo 66724fb37a25STejun Heo /* clear everything except for devices */ 66734fb37a25STejun Heo memset(link, 0, offsetof(struct ata_link, device[0])); 66744fb37a25STejun Heo 66754fb37a25STejun Heo link->ap = ap; 66768989805dSTejun Heo link->pmp = pmp; 66774fb37a25STejun Heo link->active_tag = ATA_TAG_POISON; 66784fb37a25STejun Heo link->hw_sata_spd_limit = UINT_MAX; 66794fb37a25STejun Heo 66804fb37a25STejun Heo /* can't use iterator, ap isn't initialized yet */ 66814fb37a25STejun Heo for (i = 0; i < ATA_MAX_DEVICES; i++) { 66824fb37a25STejun Heo struct ata_device *dev = &link->device[i]; 66834fb37a25STejun Heo 66844fb37a25STejun Heo dev->link = link; 66854fb37a25STejun Heo dev->devno = dev - link->device; 66864fb37a25STejun Heo ata_dev_init(dev); 66874fb37a25STejun Heo } 66884fb37a25STejun Heo } 66894fb37a25STejun Heo 66904fb37a25STejun Heo /** 66914fb37a25STejun Heo * sata_link_init_spd - Initialize link->sata_spd_limit 66924fb37a25STejun Heo * @link: Link to configure sata_spd_limit for 66934fb37a25STejun Heo * 66944fb37a25STejun Heo * Initialize @link->[hw_]sata_spd_limit to the currently 66954fb37a25STejun Heo * configured value. 66964fb37a25STejun Heo * 66974fb37a25STejun Heo * LOCKING: 66984fb37a25STejun Heo * Kernel thread context (may sleep). 66994fb37a25STejun Heo * 67004fb37a25STejun Heo * RETURNS: 67014fb37a25STejun Heo * 0 on success, -errno on failure. 67024fb37a25STejun Heo */ 6703fb7fd614STejun Heo int sata_link_init_spd(struct ata_link *link) 67044fb37a25STejun Heo { 670533267325STejun Heo u32 scontrol; 670633267325STejun Heo u8 spd; 67074fb37a25STejun Heo int rc; 67084fb37a25STejun Heo 67094fb37a25STejun Heo rc = sata_scr_read(link, SCR_CONTROL, &scontrol); 67104fb37a25STejun Heo if (rc) 67114fb37a25STejun Heo return rc; 67124fb37a25STejun Heo 67134fb37a25STejun Heo spd = (scontrol >> 4) & 0xf; 67144fb37a25STejun Heo if (spd) 67154fb37a25STejun Heo link->hw_sata_spd_limit &= (1 << spd) - 1; 67164fb37a25STejun Heo 671733267325STejun Heo ata_force_spd_limit(link); 671833267325STejun Heo 67194fb37a25STejun Heo link->sata_spd_limit = link->hw_sata_spd_limit; 67204fb37a25STejun Heo 67214fb37a25STejun Heo return 0; 67224fb37a25STejun Heo } 67234fb37a25STejun Heo 67244fb37a25STejun Heo /** 6725f3187195STejun Heo * ata_port_alloc - allocate and initialize basic ATA port resources 6726f3187195STejun Heo * @host: ATA host this allocated port belongs to 6727c6fd2807SJeff Garzik * 6728f3187195STejun Heo * Allocate and initialize basic ATA port resources. 6729f3187195STejun Heo * 6730f3187195STejun Heo * RETURNS: 6731f3187195STejun Heo * Allocate ATA port on success, NULL on failure. 6732c6fd2807SJeff Garzik * 6733c6fd2807SJeff Garzik * LOCKING: 6734f3187195STejun Heo * Inherited from calling layer (may sleep). 6735c6fd2807SJeff Garzik */ 6736f3187195STejun Heo struct ata_port *ata_port_alloc(struct ata_host *host) 6737c6fd2807SJeff Garzik { 6738f3187195STejun Heo struct ata_port *ap; 6739c6fd2807SJeff Garzik 6740f3187195STejun Heo DPRINTK("ENTER\n"); 6741f3187195STejun Heo 6742f3187195STejun Heo ap = kzalloc(sizeof(*ap), GFP_KERNEL); 6743f3187195STejun Heo if (!ap) 6744f3187195STejun Heo return NULL; 6745f3187195STejun Heo 6746f4d6d004STejun Heo ap->pflags |= ATA_PFLAG_INITIALIZING; 6747cca3974eSJeff Garzik ap->lock = &host->lock; 6748c6fd2807SJeff Garzik ap->flags = ATA_FLAG_DISABLED; 6749f3187195STejun Heo ap->print_id = -1; 6750c6fd2807SJeff Garzik ap->ctl = ATA_DEVCTL_OBS; 6751cca3974eSJeff Garzik ap->host = host; 6752f3187195STejun Heo ap->dev = host->dev; 6753c6fd2807SJeff Garzik ap->last_ctl = 0xFF; 6754c6fd2807SJeff Garzik 6755c6fd2807SJeff Garzik #if defined(ATA_VERBOSE_DEBUG) 6756c6fd2807SJeff Garzik /* turn on all debugging levels */ 6757c6fd2807SJeff Garzik ap->msg_enable = 0x00FF; 6758c6fd2807SJeff Garzik #elif defined(ATA_DEBUG) 6759c6fd2807SJeff Garzik ap->msg_enable = ATA_MSG_DRV | ATA_MSG_INFO | ATA_MSG_CTL | ATA_MSG_WARN | ATA_MSG_ERR; 6760c6fd2807SJeff Garzik #else 6761c6fd2807SJeff Garzik ap->msg_enable = ATA_MSG_DRV | ATA_MSG_ERR | ATA_MSG_WARN; 6762c6fd2807SJeff Garzik #endif 6763c6fd2807SJeff Garzik 6764442eacc3SJeff Garzik INIT_DELAYED_WORK(&ap->port_task, ata_pio_task); 676565f27f38SDavid Howells INIT_DELAYED_WORK(&ap->hotplug_task, ata_scsi_hotplug); 676665f27f38SDavid Howells INIT_WORK(&ap->scsi_rescan_task, ata_scsi_dev_rescan); 6767c6fd2807SJeff Garzik INIT_LIST_HEAD(&ap->eh_done_q); 6768c6fd2807SJeff Garzik init_waitqueue_head(&ap->eh_wait_q); 67695ddf24c5STejun Heo init_timer_deferrable(&ap->fastdrain_timer); 67705ddf24c5STejun Heo ap->fastdrain_timer.function = ata_eh_fastdrain_timerfn; 67715ddf24c5STejun Heo ap->fastdrain_timer.data = (unsigned long)ap; 6772c6fd2807SJeff Garzik 6773c6fd2807SJeff Garzik ap->cbl = ATA_CBL_NONE; 6774c6fd2807SJeff Garzik 67758989805dSTejun Heo ata_link_init(ap, &ap->link, 0); 6776c6fd2807SJeff Garzik 6777c6fd2807SJeff Garzik #ifdef ATA_IRQ_TRAP 6778c6fd2807SJeff Garzik ap->stats.unhandled_irq = 1; 6779c6fd2807SJeff Garzik ap->stats.idle_irq = 1; 6780c6fd2807SJeff Garzik #endif 6781c6fd2807SJeff Garzik return ap; 6782c6fd2807SJeff Garzik } 6783c6fd2807SJeff Garzik 6784f0d36efdSTejun Heo static void ata_host_release(struct device *gendev, void *res) 6785f0d36efdSTejun Heo { 6786f0d36efdSTejun Heo struct ata_host *host = dev_get_drvdata(gendev); 6787f0d36efdSTejun Heo int i; 6788f0d36efdSTejun Heo 6789f0d36efdSTejun Heo for (i = 0; i < host->n_ports; i++) { 6790f0d36efdSTejun Heo struct ata_port *ap = host->ports[i]; 6791f0d36efdSTejun Heo 6792ecef7253STejun Heo if (!ap) 6793ecef7253STejun Heo continue; 6794ecef7253STejun Heo 67954911487aSTejun Heo if (ap->scsi_host) 67961aa506e4STejun Heo scsi_host_put(ap->scsi_host); 67971aa506e4STejun Heo 6798633273a3STejun Heo kfree(ap->pmp_link); 67994911487aSTejun Heo kfree(ap); 68001aa506e4STejun Heo host->ports[i] = NULL; 68011aa506e4STejun Heo } 68021aa506e4STejun Heo 68031aa56ccaSTejun Heo dev_set_drvdata(gendev, NULL); 6804f0d36efdSTejun Heo } 6805f0d36efdSTejun Heo 6806c6fd2807SJeff Garzik /** 6807f3187195STejun Heo * ata_host_alloc - allocate and init basic ATA host resources 6808f3187195STejun Heo * @dev: generic device this host is associated with 6809f3187195STejun Heo * @max_ports: maximum number of ATA ports associated with this host 6810f3187195STejun Heo * 6811f3187195STejun Heo * Allocate and initialize basic ATA host resources. LLD calls 6812f3187195STejun Heo * this function to allocate a host, initializes it fully and 6813f3187195STejun Heo * attaches it using ata_host_register(). 6814f3187195STejun Heo * 6815f3187195STejun Heo * @max_ports ports are allocated and host->n_ports is 6816f3187195STejun Heo * initialized to @max_ports. The caller is allowed to decrease 6817f3187195STejun Heo * host->n_ports before calling ata_host_register(). The unused 6818f3187195STejun Heo * ports will be automatically freed on registration. 6819f3187195STejun Heo * 6820f3187195STejun Heo * RETURNS: 6821f3187195STejun Heo * Allocate ATA host on success, NULL on failure. 6822f3187195STejun Heo * 6823f3187195STejun Heo * LOCKING: 6824f3187195STejun Heo * Inherited from calling layer (may sleep). 6825f3187195STejun Heo */ 6826f3187195STejun Heo struct ata_host *ata_host_alloc(struct device *dev, int max_ports) 6827f3187195STejun Heo { 6828f3187195STejun Heo struct ata_host *host; 6829f3187195STejun Heo size_t sz; 6830f3187195STejun Heo int i; 6831f3187195STejun Heo 6832f3187195STejun Heo DPRINTK("ENTER\n"); 6833f3187195STejun Heo 6834f3187195STejun Heo if (!devres_open_group(dev, NULL, GFP_KERNEL)) 6835f3187195STejun Heo return NULL; 6836f3187195STejun Heo 6837f3187195STejun Heo /* alloc a container for our list of ATA ports (buses) */ 6838f3187195STejun Heo sz = sizeof(struct ata_host) + (max_ports + 1) * sizeof(void *); 6839f3187195STejun Heo /* alloc a container for our list of ATA ports (buses) */ 6840f3187195STejun Heo host = devres_alloc(ata_host_release, sz, GFP_KERNEL); 6841f3187195STejun Heo if (!host) 6842f3187195STejun Heo goto err_out; 6843f3187195STejun Heo 6844f3187195STejun Heo devres_add(dev, host); 6845f3187195STejun Heo dev_set_drvdata(dev, host); 6846f3187195STejun Heo 6847f3187195STejun Heo spin_lock_init(&host->lock); 6848f3187195STejun Heo host->dev = dev; 6849f3187195STejun Heo host->n_ports = max_ports; 6850f3187195STejun Heo 6851f3187195STejun Heo /* allocate ports bound to this host */ 6852f3187195STejun Heo for (i = 0; i < max_ports; i++) { 6853f3187195STejun Heo struct ata_port *ap; 6854f3187195STejun Heo 6855f3187195STejun Heo ap = ata_port_alloc(host); 6856f3187195STejun Heo if (!ap) 6857f3187195STejun Heo goto err_out; 6858f3187195STejun Heo 6859f3187195STejun Heo ap->port_no = i; 6860f3187195STejun Heo host->ports[i] = ap; 6861f3187195STejun Heo } 6862f3187195STejun Heo 6863f3187195STejun Heo devres_remove_group(dev, NULL); 6864f3187195STejun Heo return host; 6865f3187195STejun Heo 6866f3187195STejun Heo err_out: 6867f3187195STejun Heo devres_release_group(dev, NULL); 6868f3187195STejun Heo return NULL; 6869f3187195STejun Heo } 6870f3187195STejun Heo 6871f3187195STejun Heo /** 6872f5cda257STejun Heo * ata_host_alloc_pinfo - alloc host and init with port_info array 6873f5cda257STejun Heo * @dev: generic device this host is associated with 6874f5cda257STejun Heo * @ppi: array of ATA port_info to initialize host with 6875f5cda257STejun Heo * @n_ports: number of ATA ports attached to this host 6876f5cda257STejun Heo * 6877f5cda257STejun Heo * Allocate ATA host and initialize with info from @ppi. If NULL 6878f5cda257STejun Heo * terminated, @ppi may contain fewer entries than @n_ports. The 6879f5cda257STejun Heo * last entry will be used for the remaining ports. 6880f5cda257STejun Heo * 6881f5cda257STejun Heo * RETURNS: 6882f5cda257STejun Heo * Allocate ATA host on success, NULL on failure. 6883f5cda257STejun Heo * 6884f5cda257STejun Heo * LOCKING: 6885f5cda257STejun Heo * Inherited from calling layer (may sleep). 6886f5cda257STejun Heo */ 6887f5cda257STejun Heo struct ata_host *ata_host_alloc_pinfo(struct device *dev, 6888f5cda257STejun Heo const struct ata_port_info * const * ppi, 6889f5cda257STejun Heo int n_ports) 6890f5cda257STejun Heo { 6891f5cda257STejun Heo const struct ata_port_info *pi; 6892f5cda257STejun Heo struct ata_host *host; 6893f5cda257STejun Heo int i, j; 6894f5cda257STejun Heo 6895f5cda257STejun Heo host = ata_host_alloc(dev, n_ports); 6896f5cda257STejun Heo if (!host) 6897f5cda257STejun Heo return NULL; 6898f5cda257STejun Heo 6899f5cda257STejun Heo for (i = 0, j = 0, pi = NULL; i < host->n_ports; i++) { 6900f5cda257STejun Heo struct ata_port *ap = host->ports[i]; 6901f5cda257STejun Heo 6902f5cda257STejun Heo if (ppi[j]) 6903f5cda257STejun Heo pi = ppi[j++]; 6904f5cda257STejun Heo 6905f5cda257STejun Heo ap->pio_mask = pi->pio_mask; 6906f5cda257STejun Heo ap->mwdma_mask = pi->mwdma_mask; 6907f5cda257STejun Heo ap->udma_mask = pi->udma_mask; 6908f5cda257STejun Heo ap->flags |= pi->flags; 69090c88758bSTejun Heo ap->link.flags |= pi->link_flags; 6910f5cda257STejun Heo ap->ops = pi->port_ops; 6911f5cda257STejun Heo 6912f5cda257STejun Heo if (!host->ops && (pi->port_ops != &ata_dummy_port_ops)) 6913f5cda257STejun Heo host->ops = pi->port_ops; 6914f5cda257STejun Heo if (!host->private_data && pi->private_data) 6915f5cda257STejun Heo host->private_data = pi->private_data; 6916f5cda257STejun Heo } 6917f5cda257STejun Heo 6918f5cda257STejun Heo return host; 6919f5cda257STejun Heo } 6920f5cda257STejun Heo 692132ebbc0cSTejun Heo static void ata_host_stop(struct device *gendev, void *res) 692232ebbc0cSTejun Heo { 692332ebbc0cSTejun Heo struct ata_host *host = dev_get_drvdata(gendev); 692432ebbc0cSTejun Heo int i; 692532ebbc0cSTejun Heo 692632ebbc0cSTejun Heo WARN_ON(!(host->flags & ATA_HOST_STARTED)); 692732ebbc0cSTejun Heo 692832ebbc0cSTejun Heo for (i = 0; i < host->n_ports; i++) { 692932ebbc0cSTejun Heo struct ata_port *ap = host->ports[i]; 693032ebbc0cSTejun Heo 693132ebbc0cSTejun Heo if (ap->ops->port_stop) 693232ebbc0cSTejun Heo ap->ops->port_stop(ap); 693332ebbc0cSTejun Heo } 693432ebbc0cSTejun Heo 693532ebbc0cSTejun Heo if (host->ops->host_stop) 693632ebbc0cSTejun Heo host->ops->host_stop(host); 693732ebbc0cSTejun Heo } 693832ebbc0cSTejun Heo 6939f5cda257STejun Heo /** 6940ecef7253STejun Heo * ata_host_start - start and freeze ports of an ATA host 6941ecef7253STejun Heo * @host: ATA host to start ports for 6942ecef7253STejun Heo * 6943ecef7253STejun Heo * Start and then freeze ports of @host. Started status is 6944ecef7253STejun Heo * recorded in host->flags, so this function can be called 6945ecef7253STejun Heo * multiple times. Ports are guaranteed to get started only 6946f3187195STejun Heo * once. If host->ops isn't initialized yet, its set to the 6947f3187195STejun Heo * first non-dummy port ops. 6948ecef7253STejun Heo * 6949ecef7253STejun Heo * LOCKING: 6950ecef7253STejun Heo * Inherited from calling layer (may sleep). 6951ecef7253STejun Heo * 6952ecef7253STejun Heo * RETURNS: 6953ecef7253STejun Heo * 0 if all ports are started successfully, -errno otherwise. 6954ecef7253STejun Heo */ 6955ecef7253STejun Heo int ata_host_start(struct ata_host *host) 6956ecef7253STejun Heo { 695732ebbc0cSTejun Heo int have_stop = 0; 695832ebbc0cSTejun Heo void *start_dr = NULL; 6959ecef7253STejun Heo int i, rc; 6960ecef7253STejun Heo 6961ecef7253STejun Heo if (host->flags & ATA_HOST_STARTED) 6962ecef7253STejun Heo return 0; 6963ecef7253STejun Heo 6964ecef7253STejun Heo for (i = 0; i < host->n_ports; i++) { 6965ecef7253STejun Heo struct ata_port *ap = host->ports[i]; 6966ecef7253STejun Heo 6967f3187195STejun Heo if (!host->ops && !ata_port_is_dummy(ap)) 6968f3187195STejun Heo host->ops = ap->ops; 6969f3187195STejun Heo 697032ebbc0cSTejun Heo if (ap->ops->port_stop) 697132ebbc0cSTejun Heo have_stop = 1; 697232ebbc0cSTejun Heo } 697332ebbc0cSTejun Heo 697432ebbc0cSTejun Heo if (host->ops->host_stop) 697532ebbc0cSTejun Heo have_stop = 1; 697632ebbc0cSTejun Heo 697732ebbc0cSTejun Heo if (have_stop) { 697832ebbc0cSTejun Heo start_dr = devres_alloc(ata_host_stop, 0, GFP_KERNEL); 697932ebbc0cSTejun Heo if (!start_dr) 698032ebbc0cSTejun Heo return -ENOMEM; 698132ebbc0cSTejun Heo } 698232ebbc0cSTejun Heo 698332ebbc0cSTejun Heo for (i = 0; i < host->n_ports; i++) { 698432ebbc0cSTejun Heo struct ata_port *ap = host->ports[i]; 698532ebbc0cSTejun Heo 6986ecef7253STejun Heo if (ap->ops->port_start) { 6987ecef7253STejun Heo rc = ap->ops->port_start(ap); 6988ecef7253STejun Heo if (rc) { 69890f9fe9b7SAlan Cox if (rc != -ENODEV) 69900f757743SAndrew Morton dev_printk(KERN_ERR, host->dev, 69910f757743SAndrew Morton "failed to start port %d " 69920f757743SAndrew Morton "(errno=%d)\n", i, rc); 6993ecef7253STejun Heo goto err_out; 6994ecef7253STejun Heo } 6995ecef7253STejun Heo } 6996ecef7253STejun Heo ata_eh_freeze_port(ap); 6997ecef7253STejun Heo } 6998ecef7253STejun Heo 699932ebbc0cSTejun Heo if (start_dr) 700032ebbc0cSTejun Heo devres_add(host->dev, start_dr); 7001ecef7253STejun Heo host->flags |= ATA_HOST_STARTED; 7002ecef7253STejun Heo return 0; 7003ecef7253STejun Heo 7004ecef7253STejun Heo err_out: 7005ecef7253STejun Heo while (--i >= 0) { 7006ecef7253STejun Heo struct ata_port *ap = host->ports[i]; 7007ecef7253STejun Heo 7008ecef7253STejun Heo if (ap->ops->port_stop) 7009ecef7253STejun Heo ap->ops->port_stop(ap); 7010ecef7253STejun Heo } 701132ebbc0cSTejun Heo devres_free(start_dr); 7012ecef7253STejun Heo return rc; 7013ecef7253STejun Heo } 7014ecef7253STejun Heo 7015ecef7253STejun Heo /** 7016cca3974eSJeff Garzik * ata_sas_host_init - Initialize a host struct 7017cca3974eSJeff Garzik * @host: host to initialize 7018cca3974eSJeff Garzik * @dev: device host is attached to 7019cca3974eSJeff Garzik * @flags: host flags 7020c6fd2807SJeff Garzik * @ops: port_ops 7021c6fd2807SJeff Garzik * 7022c6fd2807SJeff Garzik * LOCKING: 7023c6fd2807SJeff Garzik * PCI/etc. bus probe sem. 7024c6fd2807SJeff Garzik * 7025c6fd2807SJeff Garzik */ 7026f3187195STejun Heo /* KILLME - the only user left is ipr */ 7027cca3974eSJeff Garzik void ata_host_init(struct ata_host *host, struct device *dev, 7028cca3974eSJeff Garzik unsigned long flags, const struct ata_port_operations *ops) 7029c6fd2807SJeff Garzik { 7030cca3974eSJeff Garzik spin_lock_init(&host->lock); 7031cca3974eSJeff Garzik host->dev = dev; 7032cca3974eSJeff Garzik host->flags = flags; 7033cca3974eSJeff Garzik host->ops = ops; 7034c6fd2807SJeff Garzik } 7035c6fd2807SJeff Garzik 7036c6fd2807SJeff Garzik /** 7037f3187195STejun Heo * ata_host_register - register initialized ATA host 7038f3187195STejun Heo * @host: ATA host to register 7039f3187195STejun Heo * @sht: template for SCSI host 7040c6fd2807SJeff Garzik * 7041f3187195STejun Heo * Register initialized ATA host. @host is allocated using 7042f3187195STejun Heo * ata_host_alloc() and fully initialized by LLD. This function 7043f3187195STejun Heo * starts ports, registers @host with ATA and SCSI layers and 7044f3187195STejun Heo * probe registered devices. 7045c6fd2807SJeff Garzik * 7046c6fd2807SJeff Garzik * LOCKING: 7047f3187195STejun Heo * Inherited from calling layer (may sleep). 7048c6fd2807SJeff Garzik * 7049c6fd2807SJeff Garzik * RETURNS: 7050f3187195STejun Heo * 0 on success, -errno otherwise. 7051c6fd2807SJeff Garzik */ 7052f3187195STejun Heo int ata_host_register(struct ata_host *host, struct scsi_host_template *sht) 7053c6fd2807SJeff Garzik { 7054f3187195STejun Heo int i, rc; 7055c6fd2807SJeff Garzik 7056f3187195STejun Heo /* host must have been started */ 7057f3187195STejun Heo if (!(host->flags & ATA_HOST_STARTED)) { 7058f3187195STejun Heo dev_printk(KERN_ERR, host->dev, 7059f3187195STejun Heo "BUG: trying to register unstarted host\n"); 7060f3187195STejun Heo WARN_ON(1); 7061f3187195STejun Heo return -EINVAL; 706202f076aaSAlan Cox } 7063f0d36efdSTejun Heo 7064f3187195STejun Heo /* Blow away unused ports. This happens when LLD can't 7065f3187195STejun Heo * determine the exact number of ports to allocate at 7066f3187195STejun Heo * allocation time. 7067f3187195STejun Heo */ 7068f3187195STejun Heo for (i = host->n_ports; host->ports[i]; i++) 7069f3187195STejun Heo kfree(host->ports[i]); 7070f0d36efdSTejun Heo 7071f3187195STejun Heo /* give ports names and add SCSI hosts */ 7072f3187195STejun Heo for (i = 0; i < host->n_ports; i++) 7073f3187195STejun Heo host->ports[i]->print_id = ata_print_id++; 7074c6fd2807SJeff Garzik 7075f3187195STejun Heo rc = ata_scsi_add_hosts(host, sht); 7076ecef7253STejun Heo if (rc) 7077f3187195STejun Heo return rc; 7078ecef7253STejun Heo 7079fafbae87STejun Heo /* associate with ACPI nodes */ 7080fafbae87STejun Heo ata_acpi_associate(host); 7081fafbae87STejun Heo 7082f3187195STejun Heo /* set cable, sata_spd_limit and report */ 7083cca3974eSJeff Garzik for (i = 0; i < host->n_ports; i++) { 7084cca3974eSJeff Garzik struct ata_port *ap = host->ports[i]; 7085f3187195STejun Heo unsigned long xfer_mask; 7086f3187195STejun Heo 7087f3187195STejun Heo /* set SATA cable type if still unset */ 7088f3187195STejun Heo if (ap->cbl == ATA_CBL_NONE && (ap->flags & ATA_FLAG_SATA)) 7089f3187195STejun Heo ap->cbl = ATA_CBL_SATA; 7090c6fd2807SJeff Garzik 7091c6fd2807SJeff Garzik /* init sata_spd_limit to the current value */ 70924fb37a25STejun Heo sata_link_init_spd(&ap->link); 7093c6fd2807SJeff Garzik 7094cbcdd875STejun Heo /* print per-port info to dmesg */ 7095f3187195STejun Heo xfer_mask = ata_pack_xfermask(ap->pio_mask, ap->mwdma_mask, 7096f3187195STejun Heo ap->udma_mask); 7097f3187195STejun Heo 7098abf6e8edSTejun Heo if (!ata_port_is_dummy(ap)) { 7099cbcdd875STejun Heo ata_port_printk(ap, KERN_INFO, 7100cbcdd875STejun Heo "%cATA max %s %s\n", 7101a16abc0bSTejun Heo (ap->flags & ATA_FLAG_SATA) ? 'S' : 'P', 7102f3187195STejun Heo ata_mode_string(xfer_mask), 7103cbcdd875STejun Heo ap->link.eh_info.desc); 7104abf6e8edSTejun Heo ata_ehi_clear_desc(&ap->link.eh_info); 7105abf6e8edSTejun Heo } else 7106f3187195STejun Heo ata_port_printk(ap, KERN_INFO, "DUMMY\n"); 7107c6fd2807SJeff Garzik } 7108c6fd2807SJeff Garzik 7109f3187195STejun Heo /* perform each probe synchronously */ 7110f3187195STejun Heo DPRINTK("probe begin\n"); 7111f3187195STejun Heo for (i = 0; i < host->n_ports; i++) { 7112f3187195STejun Heo struct ata_port *ap = host->ports[i]; 7113f3187195STejun Heo 7114f3187195STejun Heo /* probe */ 7115c6fd2807SJeff Garzik if (ap->ops->error_handler) { 71169af5c9c9STejun Heo struct ata_eh_info *ehi = &ap->link.eh_info; 7117c6fd2807SJeff Garzik unsigned long flags; 7118c6fd2807SJeff Garzik 7119c6fd2807SJeff Garzik ata_port_probe(ap); 7120c6fd2807SJeff Garzik 7121c6fd2807SJeff Garzik /* kick EH for boot probing */ 7122c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 7123c6fd2807SJeff Garzik 7124f58229f8STejun Heo ehi->probe_mask = 7125f58229f8STejun Heo (1 << ata_link_max_devices(&ap->link)) - 1; 7126c6fd2807SJeff Garzik ehi->action |= ATA_EH_SOFTRESET; 7127c6fd2807SJeff Garzik ehi->flags |= ATA_EHI_NO_AUTOPSY | ATA_EHI_QUIET; 7128c6fd2807SJeff Garzik 7129f4d6d004STejun Heo ap->pflags &= ~ATA_PFLAG_INITIALIZING; 7130c6fd2807SJeff Garzik ap->pflags |= ATA_PFLAG_LOADING; 7131c6fd2807SJeff Garzik ata_port_schedule_eh(ap); 7132c6fd2807SJeff Garzik 7133c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 7134c6fd2807SJeff Garzik 7135c6fd2807SJeff Garzik /* wait for EH to finish */ 7136c6fd2807SJeff Garzik ata_port_wait_eh(ap); 7137c6fd2807SJeff Garzik } else { 713844877b4eSTejun Heo DPRINTK("ata%u: bus probe begin\n", ap->print_id); 7139c6fd2807SJeff Garzik rc = ata_bus_probe(ap); 714044877b4eSTejun Heo DPRINTK("ata%u: bus probe end\n", ap->print_id); 7141c6fd2807SJeff Garzik 7142c6fd2807SJeff Garzik if (rc) { 7143c6fd2807SJeff Garzik /* FIXME: do something useful here? 7144c6fd2807SJeff Garzik * Current libata behavior will 7145c6fd2807SJeff Garzik * tear down everything when 7146c6fd2807SJeff Garzik * the module is removed 7147c6fd2807SJeff Garzik * or the h/w is unplugged. 7148c6fd2807SJeff Garzik */ 7149c6fd2807SJeff Garzik } 7150c6fd2807SJeff Garzik } 7151c6fd2807SJeff Garzik } 7152c6fd2807SJeff Garzik 7153c6fd2807SJeff Garzik /* probes are done, now scan each port's disk(s) */ 7154c6fd2807SJeff Garzik DPRINTK("host probe begin\n"); 7155cca3974eSJeff Garzik for (i = 0; i < host->n_ports; i++) { 7156cca3974eSJeff Garzik struct ata_port *ap = host->ports[i]; 7157c6fd2807SJeff Garzik 71581ae46317STejun Heo ata_scsi_scan_host(ap, 1); 7159ca77329fSKristen Carlson Accardi ata_lpm_schedule(ap, ap->pm_policy); 7160c6fd2807SJeff Garzik } 7161c6fd2807SJeff Garzik 7162f3187195STejun Heo return 0; 7163f3187195STejun Heo } 7164f3187195STejun Heo 7165f3187195STejun Heo /** 7166f5cda257STejun Heo * ata_host_activate - start host, request IRQ and register it 7167f5cda257STejun Heo * @host: target ATA host 7168f5cda257STejun Heo * @irq: IRQ to request 7169f5cda257STejun Heo * @irq_handler: irq_handler used when requesting IRQ 7170f5cda257STejun Heo * @irq_flags: irq_flags used when requesting IRQ 7171f5cda257STejun Heo * @sht: scsi_host_template to use when registering the host 7172f5cda257STejun Heo * 7173f5cda257STejun Heo * After allocating an ATA host and initializing it, most libata 7174f5cda257STejun Heo * LLDs perform three steps to activate the host - start host, 7175f5cda257STejun Heo * request IRQ and register it. This helper takes necessasry 7176f5cda257STejun Heo * arguments and performs the three steps in one go. 7177f5cda257STejun Heo * 71783d46b2e2SPaul Mundt * An invalid IRQ skips the IRQ registration and expects the host to 71793d46b2e2SPaul Mundt * have set polling mode on the port. In this case, @irq_handler 71803d46b2e2SPaul Mundt * should be NULL. 71813d46b2e2SPaul Mundt * 7182f5cda257STejun Heo * LOCKING: 7183f5cda257STejun Heo * Inherited from calling layer (may sleep). 7184f5cda257STejun Heo * 7185f5cda257STejun Heo * RETURNS: 7186f5cda257STejun Heo * 0 on success, -errno otherwise. 7187f5cda257STejun Heo */ 7188f5cda257STejun Heo int ata_host_activate(struct ata_host *host, int irq, 7189f5cda257STejun Heo irq_handler_t irq_handler, unsigned long irq_flags, 7190f5cda257STejun Heo struct scsi_host_template *sht) 7191f5cda257STejun Heo { 7192cbcdd875STejun Heo int i, rc; 7193f5cda257STejun Heo 7194f5cda257STejun Heo rc = ata_host_start(host); 7195f5cda257STejun Heo if (rc) 7196f5cda257STejun Heo return rc; 7197f5cda257STejun Heo 71983d46b2e2SPaul Mundt /* Special case for polling mode */ 71993d46b2e2SPaul Mundt if (!irq) { 72003d46b2e2SPaul Mundt WARN_ON(irq_handler); 72013d46b2e2SPaul Mundt return ata_host_register(host, sht); 72023d46b2e2SPaul Mundt } 72033d46b2e2SPaul Mundt 7204f5cda257STejun Heo rc = devm_request_irq(host->dev, irq, irq_handler, irq_flags, 7205f5cda257STejun Heo dev_driver_string(host->dev), host); 7206f5cda257STejun Heo if (rc) 7207f5cda257STejun Heo return rc; 7208f5cda257STejun Heo 7209cbcdd875STejun Heo for (i = 0; i < host->n_ports; i++) 7210cbcdd875STejun Heo ata_port_desc(host->ports[i], "irq %d", irq); 72114031826bSTejun Heo 7212f5cda257STejun Heo rc = ata_host_register(host, sht); 7213f5cda257STejun Heo /* if failed, just free the IRQ and leave ports alone */ 7214f5cda257STejun Heo if (rc) 7215f5cda257STejun Heo devm_free_irq(host->dev, irq, host); 7216f5cda257STejun Heo 7217f5cda257STejun Heo return rc; 7218f5cda257STejun Heo } 7219f5cda257STejun Heo 7220f5cda257STejun Heo /** 7221c6fd2807SJeff Garzik * ata_port_detach - Detach ATA port in prepration of device removal 7222c6fd2807SJeff Garzik * @ap: ATA port to be detached 7223c6fd2807SJeff Garzik * 7224c6fd2807SJeff Garzik * Detach all ATA devices and the associated SCSI devices of @ap; 7225c6fd2807SJeff Garzik * then, remove the associated SCSI host. @ap is guaranteed to 7226c6fd2807SJeff Garzik * be quiescent on return from this function. 7227c6fd2807SJeff Garzik * 7228c6fd2807SJeff Garzik * LOCKING: 7229c6fd2807SJeff Garzik * Kernel thread context (may sleep). 7230c6fd2807SJeff Garzik */ 7231741b7763SAdrian Bunk static void ata_port_detach(struct ata_port *ap) 7232c6fd2807SJeff Garzik { 7233c6fd2807SJeff Garzik unsigned long flags; 723441bda9c9STejun Heo struct ata_link *link; 7235f58229f8STejun Heo struct ata_device *dev; 7236c6fd2807SJeff Garzik 7237c6fd2807SJeff Garzik if (!ap->ops->error_handler) 7238c6fd2807SJeff Garzik goto skip_eh; 7239c6fd2807SJeff Garzik 7240c6fd2807SJeff Garzik /* tell EH we're leaving & flush EH */ 7241c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 7242c6fd2807SJeff Garzik ap->pflags |= ATA_PFLAG_UNLOADING; 7243c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 7244c6fd2807SJeff Garzik 7245c6fd2807SJeff Garzik ata_port_wait_eh(ap); 7246c6fd2807SJeff Garzik 72477f9ad9b8STejun Heo /* EH is now guaranteed to see UNLOADING - EH context belongs 72487f9ad9b8STejun Heo * to us. Disable all existing devices. 7249c6fd2807SJeff Garzik */ 725041bda9c9STejun Heo ata_port_for_each_link(link, ap) { 725141bda9c9STejun Heo ata_link_for_each_dev(dev, link) 7252f58229f8STejun Heo ata_dev_disable(dev); 725341bda9c9STejun Heo } 7254c6fd2807SJeff Garzik 7255c6fd2807SJeff Garzik /* Final freeze & EH. All in-flight commands are aborted. EH 7256c6fd2807SJeff Garzik * will be skipped and retrials will be terminated with bad 7257c6fd2807SJeff Garzik * target. 7258c6fd2807SJeff Garzik */ 7259c6fd2807SJeff Garzik spin_lock_irqsave(ap->lock, flags); 7260c6fd2807SJeff Garzik ata_port_freeze(ap); /* won't be thawed */ 7261c6fd2807SJeff Garzik spin_unlock_irqrestore(ap->lock, flags); 7262c6fd2807SJeff Garzik 7263c6fd2807SJeff Garzik ata_port_wait_eh(ap); 726445a66c1cSOleg Nesterov cancel_rearming_delayed_work(&ap->hotplug_task); 7265c6fd2807SJeff Garzik 7266c6fd2807SJeff Garzik skip_eh: 7267c6fd2807SJeff Garzik /* remove the associated SCSI host */ 7268cca3974eSJeff Garzik scsi_remove_host(ap->scsi_host); 7269c6fd2807SJeff Garzik } 7270c6fd2807SJeff Garzik 7271c6fd2807SJeff Garzik /** 72720529c159STejun Heo * ata_host_detach - Detach all ports of an ATA host 72730529c159STejun Heo * @host: Host to detach 72740529c159STejun Heo * 72750529c159STejun Heo * Detach all ports of @host. 72760529c159STejun Heo * 72770529c159STejun Heo * LOCKING: 72780529c159STejun Heo * Kernel thread context (may sleep). 72790529c159STejun Heo */ 72800529c159STejun Heo void ata_host_detach(struct ata_host *host) 72810529c159STejun Heo { 72820529c159STejun Heo int i; 72830529c159STejun Heo 72840529c159STejun Heo for (i = 0; i < host->n_ports; i++) 72850529c159STejun Heo ata_port_detach(host->ports[i]); 7286562f0c2dSTejun Heo 7287562f0c2dSTejun Heo /* the host is dead now, dissociate ACPI */ 7288562f0c2dSTejun Heo ata_acpi_dissociate(host); 72890529c159STejun Heo } 72900529c159STejun Heo 7291c6fd2807SJeff Garzik /** 7292c6fd2807SJeff Garzik * ata_std_ports - initialize ioaddr with standard port offsets. 7293c6fd2807SJeff Garzik * @ioaddr: IO address structure to be initialized 7294c6fd2807SJeff Garzik * 7295c6fd2807SJeff Garzik * Utility function which initializes data_addr, error_addr, 7296c6fd2807SJeff Garzik * feature_addr, nsect_addr, lbal_addr, lbam_addr, lbah_addr, 7297c6fd2807SJeff Garzik * device_addr, status_addr, and command_addr to standard offsets 7298c6fd2807SJeff Garzik * relative to cmd_addr. 7299c6fd2807SJeff Garzik * 7300c6fd2807SJeff Garzik * Does not set ctl_addr, altstatus_addr, bmdma_addr, or scr_addr. 7301c6fd2807SJeff Garzik */ 7302c6fd2807SJeff Garzik 7303c6fd2807SJeff Garzik void ata_std_ports(struct ata_ioports *ioaddr) 7304c6fd2807SJeff Garzik { 7305c6fd2807SJeff Garzik ioaddr->data_addr = ioaddr->cmd_addr + ATA_REG_DATA; 7306c6fd2807SJeff Garzik ioaddr->error_addr = ioaddr->cmd_addr + ATA_REG_ERR; 7307c6fd2807SJeff Garzik ioaddr->feature_addr = ioaddr->cmd_addr + ATA_REG_FEATURE; 7308c6fd2807SJeff Garzik ioaddr->nsect_addr = ioaddr->cmd_addr + ATA_REG_NSECT; 7309c6fd2807SJeff Garzik ioaddr->lbal_addr = ioaddr->cmd_addr + ATA_REG_LBAL; 7310c6fd2807SJeff Garzik ioaddr->lbam_addr = ioaddr->cmd_addr + ATA_REG_LBAM; 7311c6fd2807SJeff Garzik ioaddr->lbah_addr = ioaddr->cmd_addr + ATA_REG_LBAH; 7312c6fd2807SJeff Garzik ioaddr->device_addr = ioaddr->cmd_addr + ATA_REG_DEVICE; 7313c6fd2807SJeff Garzik ioaddr->status_addr = ioaddr->cmd_addr + ATA_REG_STATUS; 7314c6fd2807SJeff Garzik ioaddr->command_addr = ioaddr->cmd_addr + ATA_REG_CMD; 7315c6fd2807SJeff Garzik } 7316c6fd2807SJeff Garzik 7317c6fd2807SJeff Garzik 7318c6fd2807SJeff Garzik #ifdef CONFIG_PCI 7319c6fd2807SJeff Garzik 7320c6fd2807SJeff Garzik /** 7321c6fd2807SJeff Garzik * ata_pci_remove_one - PCI layer callback for device removal 7322c6fd2807SJeff Garzik * @pdev: PCI device that was removed 7323c6fd2807SJeff Garzik * 7324b878ca5dSTejun Heo * PCI layer indicates to libata via this hook that hot-unplug or 7325b878ca5dSTejun Heo * module unload event has occurred. Detach all ports. Resource 7326b878ca5dSTejun Heo * release is handled via devres. 7327c6fd2807SJeff Garzik * 7328c6fd2807SJeff Garzik * LOCKING: 7329c6fd2807SJeff Garzik * Inherited from PCI layer (may sleep). 7330c6fd2807SJeff Garzik */ 7331c6fd2807SJeff Garzik void ata_pci_remove_one(struct pci_dev *pdev) 7332c6fd2807SJeff Garzik { 73332855568bSJeff Garzik struct device *dev = &pdev->dev; 7334cca3974eSJeff Garzik struct ata_host *host = dev_get_drvdata(dev); 7335c6fd2807SJeff Garzik 7336f0d36efdSTejun Heo ata_host_detach(host); 7337c6fd2807SJeff Garzik } 7338c6fd2807SJeff Garzik 7339c6fd2807SJeff Garzik /* move to PCI subsystem */ 7340c6fd2807SJeff Garzik int pci_test_config_bits(struct pci_dev *pdev, const struct pci_bits *bits) 7341c6fd2807SJeff Garzik { 7342c6fd2807SJeff Garzik unsigned long tmp = 0; 7343c6fd2807SJeff Garzik 7344c6fd2807SJeff Garzik switch (bits->width) { 7345c6fd2807SJeff Garzik case 1: { 7346c6fd2807SJeff Garzik u8 tmp8 = 0; 7347c6fd2807SJeff Garzik pci_read_config_byte(pdev, bits->reg, &tmp8); 7348c6fd2807SJeff Garzik tmp = tmp8; 7349c6fd2807SJeff Garzik break; 7350c6fd2807SJeff Garzik } 7351c6fd2807SJeff Garzik case 2: { 7352c6fd2807SJeff Garzik u16 tmp16 = 0; 7353c6fd2807SJeff Garzik pci_read_config_word(pdev, bits->reg, &tmp16); 7354c6fd2807SJeff Garzik tmp = tmp16; 7355c6fd2807SJeff Garzik break; 7356c6fd2807SJeff Garzik } 7357c6fd2807SJeff Garzik case 4: { 7358c6fd2807SJeff Garzik u32 tmp32 = 0; 7359c6fd2807SJeff Garzik pci_read_config_dword(pdev, bits->reg, &tmp32); 7360c6fd2807SJeff Garzik tmp = tmp32; 7361c6fd2807SJeff Garzik break; 7362c6fd2807SJeff Garzik } 7363c6fd2807SJeff Garzik 7364c6fd2807SJeff Garzik default: 7365c6fd2807SJeff Garzik return -EINVAL; 7366c6fd2807SJeff Garzik } 7367c6fd2807SJeff Garzik 7368c6fd2807SJeff Garzik tmp &= bits->mask; 7369c6fd2807SJeff Garzik 7370c6fd2807SJeff Garzik return (tmp == bits->val) ? 1 : 0; 7371c6fd2807SJeff Garzik } 7372c6fd2807SJeff Garzik 73736ffa01d8STejun Heo #ifdef CONFIG_PM 7374c6fd2807SJeff Garzik void ata_pci_device_do_suspend(struct pci_dev *pdev, pm_message_t mesg) 7375c6fd2807SJeff Garzik { 7376c6fd2807SJeff Garzik pci_save_state(pdev); 7377c6fd2807SJeff Garzik pci_disable_device(pdev); 73784c90d971STejun Heo 73793a2d5b70SRafael J. Wysocki if (mesg.event & PM_EVENT_SLEEP) 7380c6fd2807SJeff Garzik pci_set_power_state(pdev, PCI_D3hot); 7381c6fd2807SJeff Garzik } 7382c6fd2807SJeff Garzik 7383553c4aa6STejun Heo int ata_pci_device_do_resume(struct pci_dev *pdev) 7384c6fd2807SJeff Garzik { 7385553c4aa6STejun Heo int rc; 7386553c4aa6STejun Heo 7387c6fd2807SJeff Garzik pci_set_power_state(pdev, PCI_D0); 7388c6fd2807SJeff Garzik pci_restore_state(pdev); 7389553c4aa6STejun Heo 7390f0d36efdSTejun Heo rc = pcim_enable_device(pdev); 7391553c4aa6STejun Heo if (rc) { 7392553c4aa6STejun Heo dev_printk(KERN_ERR, &pdev->dev, 7393553c4aa6STejun Heo "failed to enable device after resume (%d)\n", rc); 7394553c4aa6STejun Heo return rc; 7395553c4aa6STejun Heo } 7396553c4aa6STejun Heo 7397c6fd2807SJeff Garzik pci_set_master(pdev); 7398553c4aa6STejun Heo return 0; 7399c6fd2807SJeff Garzik } 7400c6fd2807SJeff Garzik 7401c6fd2807SJeff Garzik int ata_pci_device_suspend(struct pci_dev *pdev, pm_message_t mesg) 7402c6fd2807SJeff Garzik { 7403cca3974eSJeff Garzik struct ata_host *host = dev_get_drvdata(&pdev->dev); 7404c6fd2807SJeff Garzik int rc = 0; 7405c6fd2807SJeff Garzik 7406cca3974eSJeff Garzik rc = ata_host_suspend(host, mesg); 7407c6fd2807SJeff Garzik if (rc) 7408c6fd2807SJeff Garzik return rc; 7409c6fd2807SJeff Garzik 7410c6fd2807SJeff Garzik ata_pci_device_do_suspend(pdev, mesg); 7411c6fd2807SJeff Garzik 7412c6fd2807SJeff Garzik return 0; 7413c6fd2807SJeff Garzik } 7414c6fd2807SJeff Garzik 7415c6fd2807SJeff Garzik int ata_pci_device_resume(struct pci_dev *pdev) 7416c6fd2807SJeff Garzik { 7417cca3974eSJeff Garzik struct ata_host *host = dev_get_drvdata(&pdev->dev); 7418553c4aa6STejun Heo int rc; 7419c6fd2807SJeff Garzik 7420553c4aa6STejun Heo rc = ata_pci_device_do_resume(pdev); 7421553c4aa6STejun Heo if (rc == 0) 7422cca3974eSJeff Garzik ata_host_resume(host); 7423553c4aa6STejun Heo return rc; 7424c6fd2807SJeff Garzik } 74256ffa01d8STejun Heo #endif /* CONFIG_PM */ 74266ffa01d8STejun Heo 7427c6fd2807SJeff Garzik #endif /* CONFIG_PCI */ 7428c6fd2807SJeff Garzik 742933267325STejun Heo static int __init ata_parse_force_one(char **cur, 743033267325STejun Heo struct ata_force_ent *force_ent, 743133267325STejun Heo const char **reason) 743233267325STejun Heo { 743333267325STejun Heo /* FIXME: Currently, there's no way to tag init const data and 743433267325STejun Heo * using __initdata causes build failure on some versions of 743533267325STejun Heo * gcc. Once __initdataconst is implemented, add const to the 743633267325STejun Heo * following structure. 743733267325STejun Heo */ 743833267325STejun Heo static struct ata_force_param force_tbl[] __initdata = { 743933267325STejun Heo { "40c", .cbl = ATA_CBL_PATA40 }, 744033267325STejun Heo { "80c", .cbl = ATA_CBL_PATA80 }, 744133267325STejun Heo { "short40c", .cbl = ATA_CBL_PATA40_SHORT }, 744233267325STejun Heo { "unk", .cbl = ATA_CBL_PATA_UNK }, 744333267325STejun Heo { "ign", .cbl = ATA_CBL_PATA_IGN }, 744433267325STejun Heo { "sata", .cbl = ATA_CBL_SATA }, 744533267325STejun Heo { "1.5Gbps", .spd_limit = 1 }, 744633267325STejun Heo { "3.0Gbps", .spd_limit = 2 }, 744733267325STejun Heo { "noncq", .horkage_on = ATA_HORKAGE_NONCQ }, 744833267325STejun Heo { "ncq", .horkage_off = ATA_HORKAGE_NONCQ }, 744933267325STejun Heo { "pio0", .xfer_mask = 1 << (ATA_SHIFT_PIO + 0) }, 745033267325STejun Heo { "pio1", .xfer_mask = 1 << (ATA_SHIFT_PIO + 1) }, 745133267325STejun Heo { "pio2", .xfer_mask = 1 << (ATA_SHIFT_PIO + 2) }, 745233267325STejun Heo { "pio3", .xfer_mask = 1 << (ATA_SHIFT_PIO + 3) }, 745333267325STejun Heo { "pio4", .xfer_mask = 1 << (ATA_SHIFT_PIO + 4) }, 745433267325STejun Heo { "pio5", .xfer_mask = 1 << (ATA_SHIFT_PIO + 5) }, 745533267325STejun Heo { "pio6", .xfer_mask = 1 << (ATA_SHIFT_PIO + 6) }, 745633267325STejun Heo { "mwdma0", .xfer_mask = 1 << (ATA_SHIFT_MWDMA + 0) }, 745733267325STejun Heo { "mwdma1", .xfer_mask = 1 << (ATA_SHIFT_MWDMA + 1) }, 745833267325STejun Heo { "mwdma2", .xfer_mask = 1 << (ATA_SHIFT_MWDMA + 2) }, 745933267325STejun Heo { "mwdma3", .xfer_mask = 1 << (ATA_SHIFT_MWDMA + 3) }, 746033267325STejun Heo { "mwdma4", .xfer_mask = 1 << (ATA_SHIFT_MWDMA + 4) }, 746133267325STejun Heo { "udma0", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 0) }, 746233267325STejun Heo { "udma16", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 0) }, 746333267325STejun Heo { "udma/16", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 0) }, 746433267325STejun Heo { "udma1", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 1) }, 746533267325STejun Heo { "udma25", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 1) }, 746633267325STejun Heo { "udma/25", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 1) }, 746733267325STejun Heo { "udma2", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 2) }, 746833267325STejun Heo { "udma33", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 2) }, 746933267325STejun Heo { "udma/33", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 2) }, 747033267325STejun Heo { "udma3", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 3) }, 747133267325STejun Heo { "udma44", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 3) }, 747233267325STejun Heo { "udma/44", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 3) }, 747333267325STejun Heo { "udma4", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 4) }, 747433267325STejun Heo { "udma66", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 4) }, 747533267325STejun Heo { "udma/66", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 4) }, 747633267325STejun Heo { "udma5", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 5) }, 747733267325STejun Heo { "udma100", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 5) }, 747833267325STejun Heo { "udma/100", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 5) }, 747933267325STejun Heo { "udma6", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 6) }, 748033267325STejun Heo { "udma133", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 6) }, 748133267325STejun Heo { "udma/133", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 6) }, 748233267325STejun Heo { "udma7", .xfer_mask = 1 << (ATA_SHIFT_UDMA + 7) }, 748333267325STejun Heo }; 748433267325STejun Heo char *start = *cur, *p = *cur; 748533267325STejun Heo char *id, *val, *endp; 748633267325STejun Heo const struct ata_force_param *match_fp = NULL; 748733267325STejun Heo int nr_matches = 0, i; 748833267325STejun Heo 748933267325STejun Heo /* find where this param ends and update *cur */ 749033267325STejun Heo while (*p != '\0' && *p != ',') 749133267325STejun Heo p++; 749233267325STejun Heo 749333267325STejun Heo if (*p == '\0') 749433267325STejun Heo *cur = p; 749533267325STejun Heo else 749633267325STejun Heo *cur = p + 1; 749733267325STejun Heo 749833267325STejun Heo *p = '\0'; 749933267325STejun Heo 750033267325STejun Heo /* parse */ 750133267325STejun Heo p = strchr(start, ':'); 750233267325STejun Heo if (!p) { 750333267325STejun Heo val = strstrip(start); 750433267325STejun Heo goto parse_val; 750533267325STejun Heo } 750633267325STejun Heo *p = '\0'; 750733267325STejun Heo 750833267325STejun Heo id = strstrip(start); 750933267325STejun Heo val = strstrip(p + 1); 751033267325STejun Heo 751133267325STejun Heo /* parse id */ 751233267325STejun Heo p = strchr(id, '.'); 751333267325STejun Heo if (p) { 751433267325STejun Heo *p++ = '\0'; 751533267325STejun Heo force_ent->device = simple_strtoul(p, &endp, 10); 751633267325STejun Heo if (p == endp || *endp != '\0') { 751733267325STejun Heo *reason = "invalid device"; 751833267325STejun Heo return -EINVAL; 751933267325STejun Heo } 752033267325STejun Heo } 752133267325STejun Heo 752233267325STejun Heo force_ent->port = simple_strtoul(id, &endp, 10); 752333267325STejun Heo if (p == endp || *endp != '\0') { 752433267325STejun Heo *reason = "invalid port/link"; 752533267325STejun Heo return -EINVAL; 752633267325STejun Heo } 752733267325STejun Heo 752833267325STejun Heo parse_val: 752933267325STejun Heo /* parse val, allow shortcuts so that both 1.5 and 1.5Gbps work */ 753033267325STejun Heo for (i = 0; i < ARRAY_SIZE(force_tbl); i++) { 753133267325STejun Heo const struct ata_force_param *fp = &force_tbl[i]; 753233267325STejun Heo 753333267325STejun Heo if (strncasecmp(val, fp->name, strlen(val))) 753433267325STejun Heo continue; 753533267325STejun Heo 753633267325STejun Heo nr_matches++; 753733267325STejun Heo match_fp = fp; 753833267325STejun Heo 753933267325STejun Heo if (strcasecmp(val, fp->name) == 0) { 754033267325STejun Heo nr_matches = 1; 754133267325STejun Heo break; 754233267325STejun Heo } 754333267325STejun Heo } 754433267325STejun Heo 754533267325STejun Heo if (!nr_matches) { 754633267325STejun Heo *reason = "unknown value"; 754733267325STejun Heo return -EINVAL; 754833267325STejun Heo } 754933267325STejun Heo if (nr_matches > 1) { 755033267325STejun Heo *reason = "ambigious value"; 755133267325STejun Heo return -EINVAL; 755233267325STejun Heo } 755333267325STejun Heo 755433267325STejun Heo force_ent->param = *match_fp; 755533267325STejun Heo 755633267325STejun Heo return 0; 755733267325STejun Heo } 755833267325STejun Heo 755933267325STejun Heo static void __init ata_parse_force_param(void) 756033267325STejun Heo { 756133267325STejun Heo int idx = 0, size = 1; 756233267325STejun Heo int last_port = -1, last_device = -1; 756333267325STejun Heo char *p, *cur, *next; 756433267325STejun Heo 756533267325STejun Heo /* calculate maximum number of params and allocate force_tbl */ 756633267325STejun Heo for (p = ata_force_param_buf; *p; p++) 756733267325STejun Heo if (*p == ',') 756833267325STejun Heo size++; 756933267325STejun Heo 757033267325STejun Heo ata_force_tbl = kzalloc(sizeof(ata_force_tbl[0]) * size, GFP_KERNEL); 757133267325STejun Heo if (!ata_force_tbl) { 757233267325STejun Heo printk(KERN_WARNING "ata: failed to extend force table, " 757333267325STejun Heo "libata.force ignored\n"); 757433267325STejun Heo return; 757533267325STejun Heo } 757633267325STejun Heo 757733267325STejun Heo /* parse and populate the table */ 757833267325STejun Heo for (cur = ata_force_param_buf; *cur != '\0'; cur = next) { 757933267325STejun Heo const char *reason = ""; 758033267325STejun Heo struct ata_force_ent te = { .port = -1, .device = -1 }; 758133267325STejun Heo 758233267325STejun Heo next = cur; 758333267325STejun Heo if (ata_parse_force_one(&next, &te, &reason)) { 758433267325STejun Heo printk(KERN_WARNING "ata: failed to parse force " 758533267325STejun Heo "parameter \"%s\" (%s)\n", 758633267325STejun Heo cur, reason); 758733267325STejun Heo continue; 758833267325STejun Heo } 758933267325STejun Heo 759033267325STejun Heo if (te.port == -1) { 759133267325STejun Heo te.port = last_port; 759233267325STejun Heo te.device = last_device; 759333267325STejun Heo } 759433267325STejun Heo 759533267325STejun Heo ata_force_tbl[idx++] = te; 759633267325STejun Heo 759733267325STejun Heo last_port = te.port; 759833267325STejun Heo last_device = te.device; 759933267325STejun Heo } 760033267325STejun Heo 760133267325STejun Heo ata_force_tbl_size = idx; 760233267325STejun Heo } 7603c6fd2807SJeff Garzik 7604c6fd2807SJeff Garzik static int __init ata_init(void) 7605c6fd2807SJeff Garzik { 7606c6fd2807SJeff Garzik ata_probe_timeout *= HZ; 760733267325STejun Heo 760833267325STejun Heo ata_parse_force_param(); 760933267325STejun Heo 7610c6fd2807SJeff Garzik ata_wq = create_workqueue("ata"); 7611c6fd2807SJeff Garzik if (!ata_wq) 7612c6fd2807SJeff Garzik return -ENOMEM; 7613c6fd2807SJeff Garzik 7614c6fd2807SJeff Garzik ata_aux_wq = create_singlethread_workqueue("ata_aux"); 7615c6fd2807SJeff Garzik if (!ata_aux_wq) { 7616c6fd2807SJeff Garzik destroy_workqueue(ata_wq); 7617c6fd2807SJeff Garzik return -ENOMEM; 7618c6fd2807SJeff Garzik } 7619c6fd2807SJeff Garzik 7620c6fd2807SJeff Garzik printk(KERN_DEBUG "libata version " DRV_VERSION " loaded.\n"); 7621c6fd2807SJeff Garzik return 0; 7622c6fd2807SJeff Garzik } 7623c6fd2807SJeff Garzik 7624c6fd2807SJeff Garzik static void __exit ata_exit(void) 7625c6fd2807SJeff Garzik { 762633267325STejun Heo kfree(ata_force_tbl); 7627c6fd2807SJeff Garzik destroy_workqueue(ata_wq); 7628c6fd2807SJeff Garzik destroy_workqueue(ata_aux_wq); 7629c6fd2807SJeff Garzik } 7630c6fd2807SJeff Garzik 7631a4625085SBrian King subsys_initcall(ata_init); 7632c6fd2807SJeff Garzik module_exit(ata_exit); 7633c6fd2807SJeff Garzik 7634c6fd2807SJeff Garzik static unsigned long ratelimit_time; 7635c6fd2807SJeff Garzik static DEFINE_SPINLOCK(ata_ratelimit_lock); 7636c6fd2807SJeff Garzik 7637c6fd2807SJeff Garzik int ata_ratelimit(void) 7638c6fd2807SJeff Garzik { 7639c6fd2807SJeff Garzik int rc; 7640c6fd2807SJeff Garzik unsigned long flags; 7641c6fd2807SJeff Garzik 7642c6fd2807SJeff Garzik spin_lock_irqsave(&ata_ratelimit_lock, flags); 7643c6fd2807SJeff Garzik 7644c6fd2807SJeff Garzik if (time_after(jiffies, ratelimit_time)) { 7645c6fd2807SJeff Garzik rc = 1; 7646c6fd2807SJeff Garzik ratelimit_time = jiffies + (HZ/5); 7647c6fd2807SJeff Garzik } else 7648c6fd2807SJeff Garzik rc = 0; 7649c6fd2807SJeff Garzik 7650c6fd2807SJeff Garzik spin_unlock_irqrestore(&ata_ratelimit_lock, flags); 7651c6fd2807SJeff Garzik 7652c6fd2807SJeff Garzik return rc; 7653c6fd2807SJeff Garzik } 7654c6fd2807SJeff Garzik 7655c6fd2807SJeff Garzik /** 7656c6fd2807SJeff Garzik * ata_wait_register - wait until register value changes 7657c6fd2807SJeff Garzik * @reg: IO-mapped register 7658c6fd2807SJeff Garzik * @mask: Mask to apply to read register value 7659c6fd2807SJeff Garzik * @val: Wait condition 7660c6fd2807SJeff Garzik * @interval_msec: polling interval in milliseconds 7661c6fd2807SJeff Garzik * @timeout_msec: timeout in milliseconds 7662c6fd2807SJeff Garzik * 7663c6fd2807SJeff Garzik * Waiting for some bits of register to change is a common 7664c6fd2807SJeff Garzik * operation for ATA controllers. This function reads 32bit LE 7665c6fd2807SJeff Garzik * IO-mapped register @reg and tests for the following condition. 7666c6fd2807SJeff Garzik * 7667c6fd2807SJeff Garzik * (*@reg & mask) != val 7668c6fd2807SJeff Garzik * 7669c6fd2807SJeff Garzik * If the condition is met, it returns; otherwise, the process is 7670c6fd2807SJeff Garzik * repeated after @interval_msec until timeout. 7671c6fd2807SJeff Garzik * 7672c6fd2807SJeff Garzik * LOCKING: 7673c6fd2807SJeff Garzik * Kernel thread context (may sleep) 7674c6fd2807SJeff Garzik * 7675c6fd2807SJeff Garzik * RETURNS: 7676c6fd2807SJeff Garzik * The final register value. 7677c6fd2807SJeff Garzik */ 7678c6fd2807SJeff Garzik u32 ata_wait_register(void __iomem *reg, u32 mask, u32 val, 7679c6fd2807SJeff Garzik unsigned long interval_msec, 7680c6fd2807SJeff Garzik unsigned long timeout_msec) 7681c6fd2807SJeff Garzik { 7682c6fd2807SJeff Garzik unsigned long timeout; 7683c6fd2807SJeff Garzik u32 tmp; 7684c6fd2807SJeff Garzik 7685c6fd2807SJeff Garzik tmp = ioread32(reg); 7686c6fd2807SJeff Garzik 7687c6fd2807SJeff Garzik /* Calculate timeout _after_ the first read to make sure 7688c6fd2807SJeff Garzik * preceding writes reach the controller before starting to 7689c6fd2807SJeff Garzik * eat away the timeout. 7690c6fd2807SJeff Garzik */ 7691c6fd2807SJeff Garzik timeout = jiffies + (timeout_msec * HZ) / 1000; 7692c6fd2807SJeff Garzik 7693c6fd2807SJeff Garzik while ((tmp & mask) == val && time_before(jiffies, timeout)) { 7694c6fd2807SJeff Garzik msleep(interval_msec); 7695c6fd2807SJeff Garzik tmp = ioread32(reg); 7696c6fd2807SJeff Garzik } 7697c6fd2807SJeff Garzik 7698c6fd2807SJeff Garzik return tmp; 7699c6fd2807SJeff Garzik } 7700c6fd2807SJeff Garzik 7701c6fd2807SJeff Garzik /* 7702c6fd2807SJeff Garzik * Dummy port_ops 7703c6fd2807SJeff Garzik */ 7704c6fd2807SJeff Garzik static void ata_dummy_noret(struct ata_port *ap) { } 7705c6fd2807SJeff Garzik static int ata_dummy_ret0(struct ata_port *ap) { return 0; } 7706c6fd2807SJeff Garzik static void ata_dummy_qc_noret(struct ata_queued_cmd *qc) { } 7707c6fd2807SJeff Garzik 7708c6fd2807SJeff Garzik static u8 ata_dummy_check_status(struct ata_port *ap) 7709c6fd2807SJeff Garzik { 7710c6fd2807SJeff Garzik return ATA_DRDY; 7711c6fd2807SJeff Garzik } 7712c6fd2807SJeff Garzik 7713c6fd2807SJeff Garzik static unsigned int ata_dummy_qc_issue(struct ata_queued_cmd *qc) 7714c6fd2807SJeff Garzik { 7715c6fd2807SJeff Garzik return AC_ERR_SYSTEM; 7716c6fd2807SJeff Garzik } 7717c6fd2807SJeff Garzik 7718c6fd2807SJeff Garzik const struct ata_port_operations ata_dummy_port_ops = { 7719c6fd2807SJeff Garzik .check_status = ata_dummy_check_status, 7720c6fd2807SJeff Garzik .check_altstatus = ata_dummy_check_status, 7721c6fd2807SJeff Garzik .dev_select = ata_noop_dev_select, 7722c6fd2807SJeff Garzik .qc_prep = ata_noop_qc_prep, 7723c6fd2807SJeff Garzik .qc_issue = ata_dummy_qc_issue, 7724c6fd2807SJeff Garzik .freeze = ata_dummy_noret, 7725c6fd2807SJeff Garzik .thaw = ata_dummy_noret, 7726c6fd2807SJeff Garzik .error_handler = ata_dummy_noret, 7727c6fd2807SJeff Garzik .post_internal_cmd = ata_dummy_qc_noret, 7728c6fd2807SJeff Garzik .irq_clear = ata_dummy_noret, 7729c6fd2807SJeff Garzik .port_start = ata_dummy_ret0, 7730c6fd2807SJeff Garzik .port_stop = ata_dummy_noret, 7731c6fd2807SJeff Garzik }; 7732c6fd2807SJeff Garzik 773321b0ad4fSTejun Heo const struct ata_port_info ata_dummy_port_info = { 773421b0ad4fSTejun Heo .port_ops = &ata_dummy_port_ops, 773521b0ad4fSTejun Heo }; 773621b0ad4fSTejun Heo 7737c6fd2807SJeff Garzik /* 7738c6fd2807SJeff Garzik * libata is essentially a library of internal helper functions for 7739c6fd2807SJeff Garzik * low-level ATA host controller drivers. As such, the API/ABI is 7740c6fd2807SJeff Garzik * likely to change as new drivers are added and updated. 7741c6fd2807SJeff Garzik * Do not depend on ABI/API stability. 7742c6fd2807SJeff Garzik */ 7743c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_deb_timing_normal); 7744c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_deb_timing_hotplug); 7745c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_deb_timing_long); 7746c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_dummy_port_ops); 774721b0ad4fSTejun Heo EXPORT_SYMBOL_GPL(ata_dummy_port_info); 7748c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_std_bios_param); 7749c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_std_ports); 7750cca3974eSJeff Garzik EXPORT_SYMBOL_GPL(ata_host_init); 7751f3187195STejun Heo EXPORT_SYMBOL_GPL(ata_host_alloc); 7752f5cda257STejun Heo EXPORT_SYMBOL_GPL(ata_host_alloc_pinfo); 7753ecef7253STejun Heo EXPORT_SYMBOL_GPL(ata_host_start); 7754f3187195STejun Heo EXPORT_SYMBOL_GPL(ata_host_register); 7755f5cda257STejun Heo EXPORT_SYMBOL_GPL(ata_host_activate); 77560529c159STejun Heo EXPORT_SYMBOL_GPL(ata_host_detach); 7757c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_sg_init); 7758c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_hsm_move); 7759c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_qc_complete); 7760c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_qc_complete_multiple); 7761c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_qc_issue_prot); 7762c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_tf_load); 7763c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_tf_read); 7764c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_noop_dev_select); 7765c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_std_dev_select); 776643727fbcSJeff Garzik EXPORT_SYMBOL_GPL(sata_print_link_status); 7767c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_tf_to_fis); 7768c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_tf_from_fis); 77696357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_pack_xfermask); 77706357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_unpack_xfermask); 77716357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_xfer_mask2mode); 77726357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_xfer_mode2mask); 77736357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_xfer_mode2shift); 77746357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_mode_string); 77756357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_id_xfermask); 7776c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_check_status); 7777c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_altstatus); 7778c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_exec_command); 7779c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_port_start); 7780d92e74d3SAlan Cox EXPORT_SYMBOL_GPL(ata_sff_port_start); 7781c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_interrupt); 778204351821SAlan EXPORT_SYMBOL_GPL(ata_do_set_mode); 77830d5ff566STejun Heo EXPORT_SYMBOL_GPL(ata_data_xfer); 77840d5ff566STejun Heo EXPORT_SYMBOL_GPL(ata_data_xfer_noirq); 778531cc23b3STejun Heo EXPORT_SYMBOL_GPL(ata_std_qc_defer); 7786c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_qc_prep); 7787d26fc955SAlan Cox EXPORT_SYMBOL_GPL(ata_dumb_qc_prep); 7788c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_noop_qc_prep); 7789c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_setup); 7790c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_start); 7791c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_irq_clear); 7792c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_status); 7793c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_stop); 7794c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_freeze); 7795c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_thaw); 7796c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_drive_eh); 7797c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_error_handler); 7798c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bmdma_post_internal_cmd); 7799c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_port_probe); 780010305f0fSAlan EXPORT_SYMBOL_GPL(ata_dev_disable); 7801c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_set_spd); 7802936fd732STejun Heo EXPORT_SYMBOL_GPL(sata_link_debounce); 7803936fd732STejun Heo EXPORT_SYMBOL_GPL(sata_link_resume); 7804c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_bus_reset); 7805c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_std_prereset); 7806c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_std_softreset); 7807cc0680a5STejun Heo EXPORT_SYMBOL_GPL(sata_link_hardreset); 7808c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_std_hardreset); 7809c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_std_postreset); 7810c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_dev_classify); 7811c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_dev_pair); 7812c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_port_disable); 7813c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_ratelimit); 7814c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_wait_register); 7815c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_busy_sleep); 781688ff6eafSTejun Heo EXPORT_SYMBOL_GPL(ata_wait_after_reset); 7817d4b2bab4STejun Heo EXPORT_SYMBOL_GPL(ata_wait_ready); 7818c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_scsi_ioctl); 7819c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_scsi_queuecmd); 7820c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_scsi_slave_config); 7821c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_scsi_slave_destroy); 7822c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_scsi_change_queue_depth); 7823c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_host_intr); 7824c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_scr_valid); 7825c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_scr_read); 7826c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_scr_write); 7827c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(sata_scr_write_flush); 7828936fd732STejun Heo EXPORT_SYMBOL_GPL(ata_link_online); 7829936fd732STejun Heo EXPORT_SYMBOL_GPL(ata_link_offline); 78306ffa01d8STejun Heo #ifdef CONFIG_PM 7831cca3974eSJeff Garzik EXPORT_SYMBOL_GPL(ata_host_suspend); 7832cca3974eSJeff Garzik EXPORT_SYMBOL_GPL(ata_host_resume); 78336ffa01d8STejun Heo #endif /* CONFIG_PM */ 7834c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_id_string); 7835c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_id_c_string); 7836c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_scsi_simulate); 7837c6fd2807SJeff Garzik 7838c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pio_need_iordy); 78396357357cSTejun Heo EXPORT_SYMBOL_GPL(ata_timing_find_mode); 7840c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_timing_compute); 7841c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_timing_merge); 7842a0f79b92STejun Heo EXPORT_SYMBOL_GPL(ata_timing_cycle2mode); 7843c6fd2807SJeff Garzik 7844c6fd2807SJeff Garzik #ifdef CONFIG_PCI 7845c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(pci_test_config_bits); 7846d583bc18STejun Heo EXPORT_SYMBOL_GPL(ata_pci_init_sff_host); 78471626aeb8STejun Heo EXPORT_SYMBOL_GPL(ata_pci_init_bmdma); 7848d583bc18STejun Heo EXPORT_SYMBOL_GPL(ata_pci_prepare_sff_host); 78494e6b79faSTejun Heo EXPORT_SYMBOL_GPL(ata_pci_activate_sff_host); 7850c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_init_one); 7851c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_remove_one); 78526ffa01d8STejun Heo #ifdef CONFIG_PM 7853c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_device_do_suspend); 7854c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_device_do_resume); 7855c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_device_suspend); 7856c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_device_resume); 78576ffa01d8STejun Heo #endif /* CONFIG_PM */ 7858c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_default_filter); 7859c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_pci_clear_simplex); 7860c6fd2807SJeff Garzik #endif /* CONFIG_PCI */ 7861c6fd2807SJeff Garzik 786231f88384STejun Heo EXPORT_SYMBOL_GPL(sata_pmp_qc_defer_cmd_switch); 78633af9a77aSTejun Heo EXPORT_SYMBOL_GPL(sata_pmp_std_prereset); 78643af9a77aSTejun Heo EXPORT_SYMBOL_GPL(sata_pmp_std_hardreset); 78653af9a77aSTejun Heo EXPORT_SYMBOL_GPL(sata_pmp_std_postreset); 78663af9a77aSTejun Heo EXPORT_SYMBOL_GPL(sata_pmp_do_eh); 78673af9a77aSTejun Heo 7868b64bbc39STejun Heo EXPORT_SYMBOL_GPL(__ata_ehi_push_desc); 7869b64bbc39STejun Heo EXPORT_SYMBOL_GPL(ata_ehi_push_desc); 7870b64bbc39STejun Heo EXPORT_SYMBOL_GPL(ata_ehi_clear_desc); 7871cbcdd875STejun Heo EXPORT_SYMBOL_GPL(ata_port_desc); 7872cbcdd875STejun Heo #ifdef CONFIG_PCI 7873cbcdd875STejun Heo EXPORT_SYMBOL_GPL(ata_port_pbar_desc); 7874cbcdd875STejun Heo #endif /* CONFIG_PCI */ 7875c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_port_schedule_eh); 7876dbd82616STejun Heo EXPORT_SYMBOL_GPL(ata_link_abort); 7877c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_port_abort); 7878c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_port_freeze); 78797d77b247STejun Heo EXPORT_SYMBOL_GPL(sata_async_notification); 7880c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_eh_freeze_port); 7881c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_eh_thaw_port); 7882c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_eh_qc_complete); 7883c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_eh_qc_retry); 7884c6fd2807SJeff Garzik EXPORT_SYMBOL_GPL(ata_do_eh); 788583625006SAkira Iguchi EXPORT_SYMBOL_GPL(ata_irq_on); 7886a619f981SAkira Iguchi EXPORT_SYMBOL_GPL(ata_dev_try_classify); 7887be0d18dfSAlan Cox 7888be0d18dfSAlan Cox EXPORT_SYMBOL_GPL(ata_cable_40wire); 7889be0d18dfSAlan Cox EXPORT_SYMBOL_GPL(ata_cable_80wire); 7890be0d18dfSAlan Cox EXPORT_SYMBOL_GPL(ata_cable_unknown); 7891c88f90c3STejun Heo EXPORT_SYMBOL_GPL(ata_cable_ignore); 7892be0d18dfSAlan Cox EXPORT_SYMBOL_GPL(ata_cable_sata); 7893