xref: /openbmc/linux/drivers/ata/sata_sx4.c (revision 9aa2cba7)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *  sata_sx4.c - Promise SATA
4  *
5  *  Maintained by:  Tejun Heo <tj@kernel.org>
6  *  		    Please ALWAYS copy linux-ide@vger.kernel.org
7  *		    on emails.
8  *
9  *  Copyright 2003-2004 Red Hat, Inc.
10  *
11  *  libata documentation is available via 'make {ps|pdf}docs',
12  *  as Documentation/driver-api/libata.rst
13  *
14  *  Hardware documentation available under NDA.
15  */
16 
17 /*
18 	Theory of operation
19 	-------------------
20 
21 	The SX4 (PDC20621) chip features a single Host DMA (HDMA) copy
22 	engine, DIMM memory, and four ATA engines (one per SATA port).
23 	Data is copied to/from DIMM memory by the HDMA engine, before
24 	handing off to one (or more) of the ATA engines.  The ATA
25 	engines operate solely on DIMM memory.
26 
27 	The SX4 behaves like a PATA chip, with no SATA controls or
28 	knowledge whatsoever, leading to the presumption that
29 	PATA<->SATA bridges exist on SX4 boards, external to the
30 	PDC20621 chip itself.
31 
32 	The chip is quite capable, supporting an XOR engine and linked
33 	hardware commands (permits a string to transactions to be
34 	submitted and waited-on as a single unit), and an optional
35 	microprocessor.
36 
37 	The limiting factor is largely software.  This Linux driver was
38 	written to multiplex the single HDMA engine to copy disk
39 	transactions into a fixed DIMM memory space, from where an ATA
40 	engine takes over.  As a result, each WRITE looks like this:
41 
42 		submit HDMA packet to hardware
43 		hardware copies data from system memory to DIMM
44 		hardware raises interrupt
45 
46 		submit ATA packet to hardware
47 		hardware executes ATA WRITE command, w/ data in DIMM
48 		hardware raises interrupt
49 
50 	and each READ looks like this:
51 
52 		submit ATA packet to hardware
53 		hardware executes ATA READ command, w/ data in DIMM
54 		hardware raises interrupt
55 
56 		submit HDMA packet to hardware
57 		hardware copies data from DIMM to system memory
58 		hardware raises interrupt
59 
60 	This is a very slow, lock-step way of doing things that can
61 	certainly be improved by motivated kernel hackers.
62 
63  */
64 
65 #include <linux/kernel.h>
66 #include <linux/module.h>
67 #include <linux/pci.h>
68 #include <linux/slab.h>
69 #include <linux/blkdev.h>
70 #include <linux/delay.h>
71 #include <linux/interrupt.h>
72 #include <linux/device.h>
73 #include <scsi/scsi_host.h>
74 #include <scsi/scsi_cmnd.h>
75 #include <linux/libata.h>
76 #include "sata_promise.h"
77 
78 #define DRV_NAME	"sata_sx4"
79 #define DRV_VERSION	"0.12"
80 
81 static int dimm_test;
82 module_param(dimm_test, int, 0644);
83 MODULE_PARM_DESC(dimm_test, "Enable DIMM test during startup (1 = enabled)");
84 
85 enum {
86 	PDC_MMIO_BAR		= 3,
87 	PDC_DIMM_BAR		= 4,
88 
89 	PDC_PRD_TBL		= 0x44,	/* Direct command DMA table addr */
90 
91 	PDC_PKT_SUBMIT		= 0x40, /* Command packet pointer addr */
92 	PDC_HDMA_PKT_SUBMIT	= 0x100, /* Host DMA packet pointer addr */
93 	PDC_INT_SEQMASK		= 0x40,	/* Mask of asserted SEQ INTs */
94 	PDC_HDMA_CTLSTAT	= 0x12C, /* Host DMA control / status */
95 
96 	PDC_CTLSTAT		= 0x60,	/* IDEn control / status */
97 
98 	PDC_20621_SEQCTL	= 0x400,
99 	PDC_20621_SEQMASK	= 0x480,
100 	PDC_20621_GENERAL_CTL	= 0x484,
101 	PDC_20621_PAGE_SIZE	= (32 * 1024),
102 
103 	/* chosen, not constant, values; we design our own DIMM mem map */
104 	PDC_20621_DIMM_WINDOW	= 0x0C,	/* page# for 32K DIMM window */
105 	PDC_20621_DIMM_BASE	= 0x00200000,
106 	PDC_20621_DIMM_DATA	= (64 * 1024),
107 	PDC_DIMM_DATA_STEP	= (256 * 1024),
108 	PDC_DIMM_WINDOW_STEP	= (8 * 1024),
109 	PDC_DIMM_HOST_PRD	= (6 * 1024),
110 	PDC_DIMM_HOST_PKT	= (128 * 0),
111 	PDC_DIMM_HPKT_PRD	= (128 * 1),
112 	PDC_DIMM_ATA_PKT	= (128 * 2),
113 	PDC_DIMM_APKT_PRD	= (128 * 3),
114 	PDC_DIMM_HEADER_SZ	= PDC_DIMM_APKT_PRD + 128,
115 	PDC_PAGE_WINDOW		= 0x40,
116 	PDC_PAGE_DATA		= PDC_PAGE_WINDOW +
117 				  (PDC_20621_DIMM_DATA / PDC_20621_PAGE_SIZE),
118 	PDC_PAGE_SET		= PDC_DIMM_DATA_STEP / PDC_20621_PAGE_SIZE,
119 
120 	PDC_CHIP0_OFS		= 0xC0000, /* offset of chip #0 */
121 
122 	PDC_20621_ERR_MASK	= (1<<19) | (1<<20) | (1<<21) | (1<<22) |
123 				  (1<<23),
124 
125 	board_20621		= 0,	/* FastTrak S150 SX4 */
126 
127 	PDC_MASK_INT		= (1 << 10), /* HDMA/ATA mask int */
128 	PDC_RESET		= (1 << 11), /* HDMA/ATA reset */
129 	PDC_DMA_ENABLE		= (1 << 7),  /* DMA start/stop */
130 
131 	PDC_MAX_HDMA		= 32,
132 	PDC_HDMA_Q_MASK		= (PDC_MAX_HDMA - 1),
133 
134 	PDC_DIMM0_SPD_DEV_ADDRESS	= 0x50,
135 	PDC_DIMM1_SPD_DEV_ADDRESS	= 0x51,
136 	PDC_I2C_CONTROL			= 0x48,
137 	PDC_I2C_ADDR_DATA		= 0x4C,
138 	PDC_DIMM0_CONTROL		= 0x80,
139 	PDC_DIMM1_CONTROL		= 0x84,
140 	PDC_SDRAM_CONTROL		= 0x88,
141 	PDC_I2C_WRITE			= 0,		/* master -> slave */
142 	PDC_I2C_READ			= (1 << 6),	/* master <- slave */
143 	PDC_I2C_START			= (1 << 7),	/* start I2C proto */
144 	PDC_I2C_MASK_INT		= (1 << 5),	/* mask I2C interrupt */
145 	PDC_I2C_COMPLETE		= (1 << 16),	/* I2C normal compl. */
146 	PDC_I2C_NO_ACK			= (1 << 20),	/* slave no-ack addr */
147 	PDC_DIMM_SPD_SUBADDRESS_START	= 0x00,
148 	PDC_DIMM_SPD_SUBADDRESS_END	= 0x7F,
149 	PDC_DIMM_SPD_ROW_NUM		= 3,
150 	PDC_DIMM_SPD_COLUMN_NUM		= 4,
151 	PDC_DIMM_SPD_MODULE_ROW		= 5,
152 	PDC_DIMM_SPD_TYPE		= 11,
153 	PDC_DIMM_SPD_FRESH_RATE		= 12,
154 	PDC_DIMM_SPD_BANK_NUM		= 17,
155 	PDC_DIMM_SPD_CAS_LATENCY	= 18,
156 	PDC_DIMM_SPD_ATTRIBUTE		= 21,
157 	PDC_DIMM_SPD_ROW_PRE_CHARGE	= 27,
158 	PDC_DIMM_SPD_ROW_ACTIVE_DELAY	= 28,
159 	PDC_DIMM_SPD_RAS_CAS_DELAY	= 29,
160 	PDC_DIMM_SPD_ACTIVE_PRECHARGE	= 30,
161 	PDC_DIMM_SPD_SYSTEM_FREQ	= 126,
162 	PDC_CTL_STATUS			= 0x08,
163 	PDC_DIMM_WINDOW_CTLR		= 0x0C,
164 	PDC_TIME_CONTROL		= 0x3C,
165 	PDC_TIME_PERIOD			= 0x40,
166 	PDC_TIME_COUNTER		= 0x44,
167 	PDC_GENERAL_CTLR		= 0x484,
168 	PCI_PLL_INIT			= 0x8A531824,
169 	PCI_X_TCOUNT			= 0xEE1E5CFF,
170 
171 	/* PDC_TIME_CONTROL bits */
172 	PDC_TIMER_BUZZER		= (1 << 10),
173 	PDC_TIMER_MODE_PERIODIC		= 0,		/* bits 9:8 == 00 */
174 	PDC_TIMER_MODE_ONCE		= (1 << 8),	/* bits 9:8 == 01 */
175 	PDC_TIMER_ENABLE		= (1 << 7),
176 	PDC_TIMER_MASK_INT		= (1 << 5),
177 	PDC_TIMER_SEQ_MASK		= 0x1f,		/* SEQ ID for timer */
178 	PDC_TIMER_DEFAULT		= PDC_TIMER_MODE_ONCE |
179 					  PDC_TIMER_ENABLE |
180 					  PDC_TIMER_MASK_INT,
181 };
182 
183 #define ECC_ERASE_BUF_SZ (128 * 1024)
184 
185 struct pdc_port_priv {
186 	u8			dimm_buf[(ATA_PRD_SZ * ATA_MAX_PRD) + 512];
187 	u8			*pkt;
188 	dma_addr_t		pkt_dma;
189 };
190 
191 struct pdc_host_priv {
192 	unsigned int		doing_hdma;
193 	unsigned int		hdma_prod;
194 	unsigned int		hdma_cons;
195 	struct {
196 		struct ata_queued_cmd *qc;
197 		unsigned int	seq;
198 		unsigned long	pkt_ofs;
199 	} hdma[32];
200 };
201 
202 
203 static int pdc_sata_init_one(struct pci_dev *pdev, const struct pci_device_id *ent);
204 static void pdc_error_handler(struct ata_port *ap);
205 static void pdc_freeze(struct ata_port *ap);
206 static void pdc_thaw(struct ata_port *ap);
207 static int pdc_port_start(struct ata_port *ap);
208 static enum ata_completion_errors pdc20621_qc_prep(struct ata_queued_cmd *qc);
209 static void pdc_tf_load_mmio(struct ata_port *ap, const struct ata_taskfile *tf);
210 static void pdc_exec_command_mmio(struct ata_port *ap, const struct ata_taskfile *tf);
211 static unsigned int pdc20621_dimm_init(struct ata_host *host);
212 static int pdc20621_detect_dimm(struct ata_host *host);
213 static unsigned int pdc20621_i2c_read(struct ata_host *host,
214 				      u32 device, u32 subaddr, u32 *pdata);
215 static int pdc20621_prog_dimm0(struct ata_host *host);
216 static unsigned int pdc20621_prog_dimm_global(struct ata_host *host);
217 static void pdc20621_get_from_dimm(struct ata_host *host,
218 				   void *psource, u32 offset, u32 size);
219 static void pdc20621_put_to_dimm(struct ata_host *host,
220 				 void *psource, u32 offset, u32 size);
221 static void pdc20621_irq_clear(struct ata_port *ap);
222 static unsigned int pdc20621_qc_issue(struct ata_queued_cmd *qc);
223 static int pdc_softreset(struct ata_link *link, unsigned int *class,
224 			 unsigned long deadline);
225 static void pdc_post_internal_cmd(struct ata_queued_cmd *qc);
226 static int pdc_check_atapi_dma(struct ata_queued_cmd *qc);
227 
228 
229 static const struct scsi_host_template pdc_sata_sht = {
230 	ATA_BASE_SHT(DRV_NAME),
231 	.sg_tablesize		= LIBATA_MAX_PRD,
232 	.dma_boundary		= ATA_DMA_BOUNDARY,
233 };
234 
235 static struct ata_port_operations pdc_20621_ops = {
236 	.inherits		= &ata_sff_port_ops,
237 
238 	.check_atapi_dma	= pdc_check_atapi_dma,
239 	.qc_prep		= pdc20621_qc_prep,
240 	.qc_issue		= pdc20621_qc_issue,
241 
242 	.freeze			= pdc_freeze,
243 	.thaw			= pdc_thaw,
244 	.softreset		= pdc_softreset,
245 	.error_handler		= pdc_error_handler,
246 	.lost_interrupt		= ATA_OP_NULL,
247 	.post_internal_cmd	= pdc_post_internal_cmd,
248 
249 	.port_start		= pdc_port_start,
250 
251 	.sff_tf_load		= pdc_tf_load_mmio,
252 	.sff_exec_command	= pdc_exec_command_mmio,
253 	.sff_irq_clear		= pdc20621_irq_clear,
254 };
255 
256 static const struct ata_port_info pdc_port_info[] = {
257 	/* board_20621 */
258 	{
259 		.flags		= ATA_FLAG_SATA | ATA_FLAG_NO_ATAPI |
260 				  ATA_FLAG_PIO_POLLING,
261 		.pio_mask	= ATA_PIO4,
262 		.mwdma_mask	= ATA_MWDMA2,
263 		.udma_mask	= ATA_UDMA6,
264 		.port_ops	= &pdc_20621_ops,
265 	},
266 
267 };
268 
269 static const struct pci_device_id pdc_sata_pci_tbl[] = {
270 	{ PCI_VDEVICE(PROMISE, 0x6622), board_20621 },
271 
272 	{ }	/* terminate list */
273 };
274 
275 static struct pci_driver pdc_sata_pci_driver = {
276 	.name			= DRV_NAME,
277 	.id_table		= pdc_sata_pci_tbl,
278 	.probe			= pdc_sata_init_one,
279 	.remove			= ata_pci_remove_one,
280 };
281 
282 
283 static int pdc_port_start(struct ata_port *ap)
284 {
285 	struct device *dev = ap->host->dev;
286 	struct pdc_port_priv *pp;
287 
288 	pp = devm_kzalloc(dev, sizeof(*pp), GFP_KERNEL);
289 	if (!pp)
290 		return -ENOMEM;
291 
292 	pp->pkt = dmam_alloc_coherent(dev, 128, &pp->pkt_dma, GFP_KERNEL);
293 	if (!pp->pkt)
294 		return -ENOMEM;
295 
296 	ap->private_data = pp;
297 
298 	return 0;
299 }
300 
301 static inline void pdc20621_ata_sg(u8 *buf, unsigned int portno,
302 				   unsigned int total_len)
303 {
304 	u32 addr;
305 	unsigned int dw = PDC_DIMM_APKT_PRD >> 2;
306 	__le32 *buf32 = (__le32 *) buf;
307 
308 	/* output ATA packet S/G table */
309 	addr = PDC_20621_DIMM_BASE + PDC_20621_DIMM_DATA +
310 	       (PDC_DIMM_DATA_STEP * portno);
311 
312 	buf32[dw] = cpu_to_le32(addr);
313 	buf32[dw + 1] = cpu_to_le32(total_len | ATA_PRD_EOT);
314 }
315 
316 static inline void pdc20621_host_sg(u8 *buf, unsigned int portno,
317 				    unsigned int total_len)
318 {
319 	u32 addr;
320 	unsigned int dw = PDC_DIMM_HPKT_PRD >> 2;
321 	__le32 *buf32 = (__le32 *) buf;
322 
323 	/* output Host DMA packet S/G table */
324 	addr = PDC_20621_DIMM_BASE + PDC_20621_DIMM_DATA +
325 	       (PDC_DIMM_DATA_STEP * portno);
326 
327 	buf32[dw] = cpu_to_le32(addr);
328 	buf32[dw + 1] = cpu_to_le32(total_len | ATA_PRD_EOT);
329 }
330 
331 static inline unsigned int pdc20621_ata_pkt(struct ata_taskfile *tf,
332 					    unsigned int devno, u8 *buf,
333 					    unsigned int portno)
334 {
335 	unsigned int i, dw;
336 	__le32 *buf32 = (__le32 *) buf;
337 	u8 dev_reg;
338 
339 	unsigned int dimm_sg = PDC_20621_DIMM_BASE +
340 			       (PDC_DIMM_WINDOW_STEP * portno) +
341 			       PDC_DIMM_APKT_PRD;
342 
343 	i = PDC_DIMM_ATA_PKT;
344 
345 	/*
346 	 * Set up ATA packet
347 	 */
348 	if ((tf->protocol == ATA_PROT_DMA) && (!(tf->flags & ATA_TFLAG_WRITE)))
349 		buf[i++] = PDC_PKT_READ;
350 	else if (tf->protocol == ATA_PROT_NODATA)
351 		buf[i++] = PDC_PKT_NODATA;
352 	else
353 		buf[i++] = 0;
354 	buf[i++] = 0;			/* reserved */
355 	buf[i++] = portno + 1;		/* seq. id */
356 	buf[i++] = 0xff;		/* delay seq. id */
357 
358 	/* dimm dma S/G, and next-pkt */
359 	dw = i >> 2;
360 	if (tf->protocol == ATA_PROT_NODATA)
361 		buf32[dw] = 0;
362 	else
363 		buf32[dw] = cpu_to_le32(dimm_sg);
364 	buf32[dw + 1] = 0;
365 	i += 8;
366 
367 	if (devno == 0)
368 		dev_reg = ATA_DEVICE_OBS;
369 	else
370 		dev_reg = ATA_DEVICE_OBS | ATA_DEV1;
371 
372 	/* select device */
373 	buf[i++] = (1 << 5) | PDC_PKT_CLEAR_BSY | ATA_REG_DEVICE;
374 	buf[i++] = dev_reg;
375 
376 	/* device control register */
377 	buf[i++] = (1 << 5) | PDC_REG_DEVCTL;
378 	buf[i++] = tf->ctl;
379 
380 	return i;
381 }
382 
383 static inline void pdc20621_host_pkt(struct ata_taskfile *tf, u8 *buf,
384 				     unsigned int portno)
385 {
386 	unsigned int dw;
387 	u32 tmp;
388 	__le32 *buf32 = (__le32 *) buf;
389 
390 	unsigned int host_sg = PDC_20621_DIMM_BASE +
391 			       (PDC_DIMM_WINDOW_STEP * portno) +
392 			       PDC_DIMM_HOST_PRD;
393 	unsigned int dimm_sg = PDC_20621_DIMM_BASE +
394 			       (PDC_DIMM_WINDOW_STEP * portno) +
395 			       PDC_DIMM_HPKT_PRD;
396 
397 	dw = PDC_DIMM_HOST_PKT >> 2;
398 
399 	/*
400 	 * Set up Host DMA packet
401 	 */
402 	if ((tf->protocol == ATA_PROT_DMA) && (!(tf->flags & ATA_TFLAG_WRITE)))
403 		tmp = PDC_PKT_READ;
404 	else
405 		tmp = 0;
406 	tmp |= ((portno + 1 + 4) << 16);	/* seq. id */
407 	tmp |= (0xff << 24);			/* delay seq. id */
408 	buf32[dw + 0] = cpu_to_le32(tmp);
409 	buf32[dw + 1] = cpu_to_le32(host_sg);
410 	buf32[dw + 2] = cpu_to_le32(dimm_sg);
411 	buf32[dw + 3] = 0;
412 }
413 
414 static void pdc20621_dma_prep(struct ata_queued_cmd *qc)
415 {
416 	struct scatterlist *sg;
417 	struct ata_port *ap = qc->ap;
418 	struct pdc_port_priv *pp = ap->private_data;
419 	void __iomem *mmio = ap->host->iomap[PDC_MMIO_BAR];
420 	void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR];
421 	unsigned int portno = ap->port_no;
422 	unsigned int i, si, idx, total_len = 0, sgt_len;
423 	__le32 *buf = (__le32 *) &pp->dimm_buf[PDC_DIMM_HEADER_SZ];
424 
425 	WARN_ON(!(qc->flags & ATA_QCFLAG_DMAMAP));
426 
427 	/* hard-code chip #0 */
428 	mmio += PDC_CHIP0_OFS;
429 
430 	/*
431 	 * Build S/G table
432 	 */
433 	idx = 0;
434 	for_each_sg(qc->sg, sg, qc->n_elem, si) {
435 		buf[idx++] = cpu_to_le32(sg_dma_address(sg));
436 		buf[idx++] = cpu_to_le32(sg_dma_len(sg));
437 		total_len += sg_dma_len(sg);
438 	}
439 	buf[idx - 1] |= cpu_to_le32(ATA_PRD_EOT);
440 	sgt_len = idx * 4;
441 
442 	/*
443 	 * Build ATA, host DMA packets
444 	 */
445 	pdc20621_host_sg(&pp->dimm_buf[0], portno, total_len);
446 	pdc20621_host_pkt(&qc->tf, &pp->dimm_buf[0], portno);
447 
448 	pdc20621_ata_sg(&pp->dimm_buf[0], portno, total_len);
449 	i = pdc20621_ata_pkt(&qc->tf, qc->dev->devno, &pp->dimm_buf[0], portno);
450 
451 	if (qc->tf.flags & ATA_TFLAG_LBA48)
452 		i = pdc_prep_lba48(&qc->tf, &pp->dimm_buf[0], i);
453 	else
454 		i = pdc_prep_lba28(&qc->tf, &pp->dimm_buf[0], i);
455 
456 	pdc_pkt_footer(&qc->tf, &pp->dimm_buf[0], i);
457 
458 	/* copy three S/G tables and two packets to DIMM MMIO window */
459 	memcpy_toio(dimm_mmio + (portno * PDC_DIMM_WINDOW_STEP),
460 		    &pp->dimm_buf, PDC_DIMM_HEADER_SZ);
461 	memcpy_toio(dimm_mmio + (portno * PDC_DIMM_WINDOW_STEP) +
462 		    PDC_DIMM_HOST_PRD,
463 		    &pp->dimm_buf[PDC_DIMM_HEADER_SZ], sgt_len);
464 
465 	/* force host FIFO dump */
466 	writel(0x00000001, mmio + PDC_20621_GENERAL_CTL);
467 
468 	readl(dimm_mmio);	/* MMIO PCI posting flush */
469 
470 	ata_port_dbg(ap, "ata pkt buf ofs %u, prd size %u, mmio copied\n",
471 		     i, sgt_len);
472 }
473 
474 static void pdc20621_nodata_prep(struct ata_queued_cmd *qc)
475 {
476 	struct ata_port *ap = qc->ap;
477 	struct pdc_port_priv *pp = ap->private_data;
478 	void __iomem *mmio = ap->host->iomap[PDC_MMIO_BAR];
479 	void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR];
480 	unsigned int portno = ap->port_no;
481 	unsigned int i;
482 
483 	/* hard-code chip #0 */
484 	mmio += PDC_CHIP0_OFS;
485 
486 	i = pdc20621_ata_pkt(&qc->tf, qc->dev->devno, &pp->dimm_buf[0], portno);
487 
488 	if (qc->tf.flags & ATA_TFLAG_LBA48)
489 		i = pdc_prep_lba48(&qc->tf, &pp->dimm_buf[0], i);
490 	else
491 		i = pdc_prep_lba28(&qc->tf, &pp->dimm_buf[0], i);
492 
493 	pdc_pkt_footer(&qc->tf, &pp->dimm_buf[0], i);
494 
495 	/* copy three S/G tables and two packets to DIMM MMIO window */
496 	memcpy_toio(dimm_mmio + (portno * PDC_DIMM_WINDOW_STEP),
497 		    &pp->dimm_buf, PDC_DIMM_HEADER_SZ);
498 
499 	/* force host FIFO dump */
500 	writel(0x00000001, mmio + PDC_20621_GENERAL_CTL);
501 
502 	readl(dimm_mmio);	/* MMIO PCI posting flush */
503 
504 	ata_port_dbg(ap, "ata pkt buf ofs %u, mmio copied\n", i);
505 }
506 
507 static enum ata_completion_errors pdc20621_qc_prep(struct ata_queued_cmd *qc)
508 {
509 	switch (qc->tf.protocol) {
510 	case ATA_PROT_DMA:
511 		pdc20621_dma_prep(qc);
512 		break;
513 	case ATA_PROT_NODATA:
514 		pdc20621_nodata_prep(qc);
515 		break;
516 	default:
517 		break;
518 	}
519 
520 	return AC_ERR_OK;
521 }
522 
523 static void __pdc20621_push_hdma(struct ata_queued_cmd *qc,
524 				 unsigned int seq,
525 				 u32 pkt_ofs)
526 {
527 	struct ata_port *ap = qc->ap;
528 	struct ata_host *host = ap->host;
529 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
530 
531 	/* hard-code chip #0 */
532 	mmio += PDC_CHIP0_OFS;
533 
534 	writel(0x00000001, mmio + PDC_20621_SEQCTL + (seq * 4));
535 	readl(mmio + PDC_20621_SEQCTL + (seq * 4));	/* flush */
536 
537 	writel(pkt_ofs, mmio + PDC_HDMA_PKT_SUBMIT);
538 	readl(mmio + PDC_HDMA_PKT_SUBMIT);	/* flush */
539 }
540 
541 static void pdc20621_push_hdma(struct ata_queued_cmd *qc,
542 				unsigned int seq,
543 				u32 pkt_ofs)
544 {
545 	struct ata_port *ap = qc->ap;
546 	struct pdc_host_priv *pp = ap->host->private_data;
547 	unsigned int idx = pp->hdma_prod & PDC_HDMA_Q_MASK;
548 
549 	if (!pp->doing_hdma) {
550 		__pdc20621_push_hdma(qc, seq, pkt_ofs);
551 		pp->doing_hdma = 1;
552 		return;
553 	}
554 
555 	pp->hdma[idx].qc = qc;
556 	pp->hdma[idx].seq = seq;
557 	pp->hdma[idx].pkt_ofs = pkt_ofs;
558 	pp->hdma_prod++;
559 }
560 
561 static void pdc20621_pop_hdma(struct ata_queued_cmd *qc)
562 {
563 	struct ata_port *ap = qc->ap;
564 	struct pdc_host_priv *pp = ap->host->private_data;
565 	unsigned int idx = pp->hdma_cons & PDC_HDMA_Q_MASK;
566 
567 	/* if nothing on queue, we're done */
568 	if (pp->hdma_prod == pp->hdma_cons) {
569 		pp->doing_hdma = 0;
570 		return;
571 	}
572 
573 	__pdc20621_push_hdma(pp->hdma[idx].qc, pp->hdma[idx].seq,
574 			     pp->hdma[idx].pkt_ofs);
575 	pp->hdma_cons++;
576 }
577 
578 static void pdc20621_dump_hdma(struct ata_queued_cmd *qc)
579 {
580 	struct ata_port *ap = qc->ap;
581 	unsigned int port_no = ap->port_no;
582 	void __iomem *dimm_mmio = ap->host->iomap[PDC_DIMM_BAR];
583 
584 	dimm_mmio += (port_no * PDC_DIMM_WINDOW_STEP);
585 	dimm_mmio += PDC_DIMM_HOST_PKT;
586 
587 	ata_port_dbg(ap, "HDMA 0x%08X 0x%08X 0x%08X 0x%08X\n",
588 		     readl(dimm_mmio), readl(dimm_mmio + 4),
589 		     readl(dimm_mmio + 8), readl(dimm_mmio + 12));
590 }
591 
592 static void pdc20621_packet_start(struct ata_queued_cmd *qc)
593 {
594 	struct ata_port *ap = qc->ap;
595 	struct ata_host *host = ap->host;
596 	unsigned int port_no = ap->port_no;
597 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
598 	unsigned int rw = (qc->tf.flags & ATA_TFLAG_WRITE);
599 	u8 seq = (u8) (port_no + 1);
600 	unsigned int port_ofs;
601 
602 	/* hard-code chip #0 */
603 	mmio += PDC_CHIP0_OFS;
604 
605 	wmb();			/* flush PRD, pkt writes */
606 
607 	port_ofs = PDC_20621_DIMM_BASE + (PDC_DIMM_WINDOW_STEP * port_no);
608 
609 	/* if writing, we (1) DMA to DIMM, then (2) do ATA command */
610 	if (rw && qc->tf.protocol == ATA_PROT_DMA) {
611 		seq += 4;
612 
613 		pdc20621_dump_hdma(qc);
614 		pdc20621_push_hdma(qc, seq, port_ofs + PDC_DIMM_HOST_PKT);
615 		ata_port_dbg(ap, "queued ofs 0x%x (%u), seq %u\n",
616 			port_ofs + PDC_DIMM_HOST_PKT,
617 			port_ofs + PDC_DIMM_HOST_PKT,
618 			seq);
619 	} else {
620 		writel(0x00000001, mmio + PDC_20621_SEQCTL + (seq * 4));
621 		readl(mmio + PDC_20621_SEQCTL + (seq * 4));	/* flush */
622 
623 		writel(port_ofs + PDC_DIMM_ATA_PKT,
624 		       ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT);
625 		readl(ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT);
626 		ata_port_dbg(ap, "submitted ofs 0x%x (%u), seq %u\n",
627 			port_ofs + PDC_DIMM_ATA_PKT,
628 			port_ofs + PDC_DIMM_ATA_PKT,
629 			seq);
630 	}
631 }
632 
633 static unsigned int pdc20621_qc_issue(struct ata_queued_cmd *qc)
634 {
635 	switch (qc->tf.protocol) {
636 	case ATA_PROT_NODATA:
637 		if (qc->tf.flags & ATA_TFLAG_POLLING)
638 			break;
639 		fallthrough;
640 	case ATA_PROT_DMA:
641 		pdc20621_packet_start(qc);
642 		return 0;
643 
644 	case ATAPI_PROT_DMA:
645 		BUG();
646 		break;
647 
648 	default:
649 		break;
650 	}
651 
652 	return ata_sff_qc_issue(qc);
653 }
654 
655 static inline unsigned int pdc20621_host_intr(struct ata_port *ap,
656 					  struct ata_queued_cmd *qc,
657 					  unsigned int doing_hdma,
658 					  void __iomem *mmio)
659 {
660 	unsigned int port_no = ap->port_no;
661 	unsigned int port_ofs =
662 		PDC_20621_DIMM_BASE + (PDC_DIMM_WINDOW_STEP * port_no);
663 	u8 status;
664 	unsigned int handled = 0;
665 
666 	if ((qc->tf.protocol == ATA_PROT_DMA) &&	/* read */
667 	    (!(qc->tf.flags & ATA_TFLAG_WRITE))) {
668 
669 		/* step two - DMA from DIMM to host */
670 		if (doing_hdma) {
671 			ata_port_dbg(ap, "read hdma, 0x%x 0x%x\n",
672 				readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT));
673 			/* get drive status; clear intr; complete txn */
674 			qc->err_mask |= ac_err_mask(ata_wait_idle(ap));
675 			ata_qc_complete(qc);
676 			pdc20621_pop_hdma(qc);
677 		}
678 
679 		/* step one - exec ATA command */
680 		else {
681 			u8 seq = (u8) (port_no + 1 + 4);
682 			ata_port_dbg(ap, "read ata, 0x%x 0x%x\n",
683 				readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT));
684 
685 			/* submit hdma pkt */
686 			pdc20621_dump_hdma(qc);
687 			pdc20621_push_hdma(qc, seq,
688 					   port_ofs + PDC_DIMM_HOST_PKT);
689 		}
690 		handled = 1;
691 
692 	} else if (qc->tf.protocol == ATA_PROT_DMA) {	/* write */
693 
694 		/* step one - DMA from host to DIMM */
695 		if (doing_hdma) {
696 			u8 seq = (u8) (port_no + 1);
697 			ata_port_dbg(ap, "write hdma, 0x%x 0x%x\n",
698 				readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT));
699 
700 			/* submit ata pkt */
701 			writel(0x00000001, mmio + PDC_20621_SEQCTL + (seq * 4));
702 			readl(mmio + PDC_20621_SEQCTL + (seq * 4));
703 			writel(port_ofs + PDC_DIMM_ATA_PKT,
704 			       ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT);
705 			readl(ap->ioaddr.cmd_addr + PDC_PKT_SUBMIT);
706 		}
707 
708 		/* step two - execute ATA command */
709 		else {
710 			ata_port_dbg(ap, "write ata, 0x%x 0x%x\n",
711 				readl(mmio + 0x104), readl(mmio + PDC_HDMA_CTLSTAT));
712 			/* get drive status; clear intr; complete txn */
713 			qc->err_mask |= ac_err_mask(ata_wait_idle(ap));
714 			ata_qc_complete(qc);
715 			pdc20621_pop_hdma(qc);
716 		}
717 		handled = 1;
718 
719 	/* command completion, but no data xfer */
720 	} else if (qc->tf.protocol == ATA_PROT_NODATA) {
721 
722 		status = ata_sff_busy_wait(ap, ATA_BUSY | ATA_DRQ, 1000);
723 		ata_port_dbg(ap, "BUS_NODATA (drv_stat 0x%X)\n", status);
724 		qc->err_mask |= ac_err_mask(status);
725 		ata_qc_complete(qc);
726 		handled = 1;
727 
728 	} else {
729 		ap->stats.idle_irq++;
730 	}
731 
732 	return handled;
733 }
734 
735 static void pdc20621_irq_clear(struct ata_port *ap)
736 {
737 	ioread8(ap->ioaddr.status_addr);
738 }
739 
740 static irqreturn_t pdc20621_interrupt(int irq, void *dev_instance)
741 {
742 	struct ata_host *host = dev_instance;
743 	struct ata_port *ap;
744 	u32 mask = 0;
745 	unsigned int i, tmp, port_no;
746 	unsigned int handled = 0;
747 	void __iomem *mmio_base;
748 
749 	if (!host || !host->iomap[PDC_MMIO_BAR])
750 		return IRQ_NONE;
751 
752 	mmio_base = host->iomap[PDC_MMIO_BAR];
753 
754 	/* reading should also clear interrupts */
755 	mmio_base += PDC_CHIP0_OFS;
756 	mask = readl(mmio_base + PDC_20621_SEQMASK);
757 
758 	if (mask == 0xffffffff)
759 		return IRQ_NONE;
760 
761 	mask &= 0xffff;		/* only 16 tags possible */
762 	if (!mask)
763 		return IRQ_NONE;
764 
765 	spin_lock(&host->lock);
766 
767 	for (i = 1; i < 9; i++) {
768 		port_no = i - 1;
769 		if (port_no > 3)
770 			port_no -= 4;
771 		if (port_no >= host->n_ports)
772 			ap = NULL;
773 		else
774 			ap = host->ports[port_no];
775 		tmp = mask & (1 << i);
776 		if (ap)
777 			ata_port_dbg(ap, "seq %u, tmp %x\n", i, tmp);
778 		if (tmp && ap) {
779 			struct ata_queued_cmd *qc;
780 
781 			qc = ata_qc_from_tag(ap, ap->link.active_tag);
782 			if (qc && (!(qc->tf.flags & ATA_TFLAG_POLLING)))
783 				handled += pdc20621_host_intr(ap, qc, (i > 4),
784 							      mmio_base);
785 		}
786 	}
787 
788 	spin_unlock(&host->lock);
789 
790 	return IRQ_RETVAL(handled);
791 }
792 
793 static void pdc_freeze(struct ata_port *ap)
794 {
795 	void __iomem *mmio = ap->ioaddr.cmd_addr;
796 	u32 tmp;
797 
798 	/* FIXME: if all 4 ATA engines are stopped, also stop HDMA engine */
799 
800 	tmp = readl(mmio + PDC_CTLSTAT);
801 	tmp |= PDC_MASK_INT;
802 	tmp &= ~PDC_DMA_ENABLE;
803 	writel(tmp, mmio + PDC_CTLSTAT);
804 	readl(mmio + PDC_CTLSTAT); /* flush */
805 }
806 
807 static void pdc_thaw(struct ata_port *ap)
808 {
809 	void __iomem *mmio = ap->ioaddr.cmd_addr;
810 	u32 tmp;
811 
812 	/* FIXME: start HDMA engine, if zero ATA engines running */
813 
814 	/* clear IRQ */
815 	ioread8(ap->ioaddr.status_addr);
816 
817 	/* turn IRQ back on */
818 	tmp = readl(mmio + PDC_CTLSTAT);
819 	tmp &= ~PDC_MASK_INT;
820 	writel(tmp, mmio + PDC_CTLSTAT);
821 	readl(mmio + PDC_CTLSTAT); /* flush */
822 }
823 
824 static void pdc_reset_port(struct ata_port *ap)
825 {
826 	void __iomem *mmio = ap->ioaddr.cmd_addr + PDC_CTLSTAT;
827 	unsigned int i;
828 	u32 tmp;
829 
830 	/* FIXME: handle HDMA copy engine */
831 
832 	for (i = 11; i > 0; i--) {
833 		tmp = readl(mmio);
834 		if (tmp & PDC_RESET)
835 			break;
836 
837 		udelay(100);
838 
839 		tmp |= PDC_RESET;
840 		writel(tmp, mmio);
841 	}
842 
843 	tmp &= ~PDC_RESET;
844 	writel(tmp, mmio);
845 	readl(mmio);	/* flush */
846 }
847 
848 static int pdc_softreset(struct ata_link *link, unsigned int *class,
849 			 unsigned long deadline)
850 {
851 	pdc_reset_port(link->ap);
852 	return ata_sff_softreset(link, class, deadline);
853 }
854 
855 static void pdc_error_handler(struct ata_port *ap)
856 {
857 	if (!ata_port_is_frozen(ap))
858 		pdc_reset_port(ap);
859 
860 	ata_sff_error_handler(ap);
861 }
862 
863 static void pdc_post_internal_cmd(struct ata_queued_cmd *qc)
864 {
865 	struct ata_port *ap = qc->ap;
866 
867 	/* make DMA engine forget about the failed command */
868 	if (qc->flags & ATA_QCFLAG_EH)
869 		pdc_reset_port(ap);
870 }
871 
872 static int pdc_check_atapi_dma(struct ata_queued_cmd *qc)
873 {
874 	u8 *scsicmd = qc->scsicmd->cmnd;
875 	int pio = 1; /* atapi dma off by default */
876 
877 	/* Whitelist commands that may use DMA. */
878 	switch (scsicmd[0]) {
879 	case WRITE_12:
880 	case WRITE_10:
881 	case WRITE_6:
882 	case READ_12:
883 	case READ_10:
884 	case READ_6:
885 	case 0xad: /* READ_DVD_STRUCTURE */
886 	case 0xbe: /* READ_CD */
887 		pio = 0;
888 	}
889 	/* -45150 (FFFF4FA2) to -1 (FFFFFFFF) shall use PIO mode */
890 	if (scsicmd[0] == WRITE_10) {
891 		unsigned int lba =
892 			(scsicmd[2] << 24) |
893 			(scsicmd[3] << 16) |
894 			(scsicmd[4] << 8) |
895 			scsicmd[5];
896 		if (lba >= 0xFFFF4FA2)
897 			pio = 1;
898 	}
899 	return pio;
900 }
901 
902 static void pdc_tf_load_mmio(struct ata_port *ap, const struct ata_taskfile *tf)
903 {
904 	WARN_ON(tf->protocol == ATA_PROT_DMA ||
905 		tf->protocol == ATAPI_PROT_DMA);
906 	ata_sff_tf_load(ap, tf);
907 }
908 
909 
910 static void pdc_exec_command_mmio(struct ata_port *ap, const struct ata_taskfile *tf)
911 {
912 	WARN_ON(tf->protocol == ATA_PROT_DMA ||
913 		tf->protocol == ATAPI_PROT_DMA);
914 	ata_sff_exec_command(ap, tf);
915 }
916 
917 
918 static void pdc_sata_setup_port(struct ata_ioports *port, void __iomem *base)
919 {
920 	port->cmd_addr		= base;
921 	port->data_addr		= base;
922 	port->feature_addr	=
923 	port->error_addr	= base + 0x4;
924 	port->nsect_addr	= base + 0x8;
925 	port->lbal_addr		= base + 0xc;
926 	port->lbam_addr		= base + 0x10;
927 	port->lbah_addr		= base + 0x14;
928 	port->device_addr	= base + 0x18;
929 	port->command_addr	=
930 	port->status_addr	= base + 0x1c;
931 	port->altstatus_addr	=
932 	port->ctl_addr		= base + 0x38;
933 }
934 
935 
936 static void pdc20621_get_from_dimm(struct ata_host *host, void *psource,
937 				   u32 offset, u32 size)
938 {
939 	u32 window_size;
940 	u16 idx;
941 	u8 page_mask;
942 	long dist;
943 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
944 	void __iomem *dimm_mmio = host->iomap[PDC_DIMM_BAR];
945 
946 	/* hard-code chip #0 */
947 	mmio += PDC_CHIP0_OFS;
948 
949 	page_mask = 0x00;
950 	window_size = 0x2000 * 4; /* 32K byte uchar size */
951 	idx = (u16) (offset / window_size);
952 
953 	writel(0x01, mmio + PDC_GENERAL_CTLR);
954 	readl(mmio + PDC_GENERAL_CTLR);
955 	writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR);
956 	readl(mmio + PDC_DIMM_WINDOW_CTLR);
957 
958 	offset -= (idx * window_size);
959 	idx++;
960 	dist = ((long) (window_size - (offset + size))) >= 0 ? size :
961 		(long) (window_size - offset);
962 	memcpy_fromio(psource, dimm_mmio + offset / 4, dist);
963 
964 	psource += dist;
965 	size -= dist;
966 	for (; (long) size >= (long) window_size ;) {
967 		writel(0x01, mmio + PDC_GENERAL_CTLR);
968 		readl(mmio + PDC_GENERAL_CTLR);
969 		writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR);
970 		readl(mmio + PDC_DIMM_WINDOW_CTLR);
971 		memcpy_fromio(psource, dimm_mmio, window_size / 4);
972 		psource += window_size;
973 		size -= window_size;
974 		idx++;
975 	}
976 
977 	if (size) {
978 		writel(0x01, mmio + PDC_GENERAL_CTLR);
979 		readl(mmio + PDC_GENERAL_CTLR);
980 		writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR);
981 		readl(mmio + PDC_DIMM_WINDOW_CTLR);
982 		memcpy_fromio(psource, dimm_mmio, size / 4);
983 	}
984 }
985 
986 
987 static void pdc20621_put_to_dimm(struct ata_host *host, void *psource,
988 				 u32 offset, u32 size)
989 {
990 	u32 window_size;
991 	u16 idx;
992 	u8 page_mask;
993 	long dist;
994 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
995 	void __iomem *dimm_mmio = host->iomap[PDC_DIMM_BAR];
996 
997 	/* hard-code chip #0 */
998 	mmio += PDC_CHIP0_OFS;
999 
1000 	page_mask = 0x00;
1001 	window_size = 0x2000 * 4;       /* 32K byte uchar size */
1002 	idx = (u16) (offset / window_size);
1003 
1004 	writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR);
1005 	readl(mmio + PDC_DIMM_WINDOW_CTLR);
1006 	offset -= (idx * window_size);
1007 	idx++;
1008 	dist = ((long)(s32)(window_size - (offset + size))) >= 0 ? size :
1009 		(long) (window_size - offset);
1010 	memcpy_toio(dimm_mmio + offset / 4, psource, dist);
1011 	writel(0x01, mmio + PDC_GENERAL_CTLR);
1012 	readl(mmio + PDC_GENERAL_CTLR);
1013 
1014 	psource += dist;
1015 	size -= dist;
1016 	for (; (long) size >= (long) window_size ;) {
1017 		writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR);
1018 		readl(mmio + PDC_DIMM_WINDOW_CTLR);
1019 		memcpy_toio(dimm_mmio, psource, window_size / 4);
1020 		writel(0x01, mmio + PDC_GENERAL_CTLR);
1021 		readl(mmio + PDC_GENERAL_CTLR);
1022 		psource += window_size;
1023 		size -= window_size;
1024 		idx++;
1025 	}
1026 
1027 	if (size) {
1028 		writel(((idx) << page_mask), mmio + PDC_DIMM_WINDOW_CTLR);
1029 		readl(mmio + PDC_DIMM_WINDOW_CTLR);
1030 		memcpy_toio(dimm_mmio, psource, size / 4);
1031 		writel(0x01, mmio + PDC_GENERAL_CTLR);
1032 		readl(mmio + PDC_GENERAL_CTLR);
1033 	}
1034 }
1035 
1036 
1037 static unsigned int pdc20621_i2c_read(struct ata_host *host, u32 device,
1038 				      u32 subaddr, u32 *pdata)
1039 {
1040 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
1041 	u32 i2creg  = 0;
1042 	u32 status;
1043 	u32 count = 0;
1044 
1045 	/* hard-code chip #0 */
1046 	mmio += PDC_CHIP0_OFS;
1047 
1048 	i2creg |= device << 24;
1049 	i2creg |= subaddr << 16;
1050 
1051 	/* Set the device and subaddress */
1052 	writel(i2creg, mmio + PDC_I2C_ADDR_DATA);
1053 	readl(mmio + PDC_I2C_ADDR_DATA);
1054 
1055 	/* Write Control to perform read operation, mask int */
1056 	writel(PDC_I2C_READ | PDC_I2C_START | PDC_I2C_MASK_INT,
1057 	       mmio + PDC_I2C_CONTROL);
1058 
1059 	for (count = 0; count <= 1000; count ++) {
1060 		status = readl(mmio + PDC_I2C_CONTROL);
1061 		if (status & PDC_I2C_COMPLETE) {
1062 			status = readl(mmio + PDC_I2C_ADDR_DATA);
1063 			break;
1064 		} else if (count == 1000)
1065 			return 0;
1066 	}
1067 
1068 	*pdata = (status >> 8) & 0x000000ff;
1069 	return 1;
1070 }
1071 
1072 
1073 static int pdc20621_detect_dimm(struct ata_host *host)
1074 {
1075 	u32 data = 0;
1076 	if (pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS,
1077 			     PDC_DIMM_SPD_SYSTEM_FREQ, &data)) {
1078 		if (data == 100)
1079 			return 100;
1080 	} else
1081 		return 0;
1082 
1083 	if (pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS, 9, &data)) {
1084 		if (data <= 0x75)
1085 			return 133;
1086 	} else
1087 		return 0;
1088 
1089 	return 0;
1090 }
1091 
1092 
1093 static int pdc20621_prog_dimm0(struct ata_host *host)
1094 {
1095 	u32 spd0[50];
1096 	u32 data = 0;
1097 	int size, i;
1098 	u8 bdimmsize;
1099 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
1100 	static const struct {
1101 		unsigned int reg;
1102 		unsigned int ofs;
1103 	} pdc_i2c_read_data [] = {
1104 		{ PDC_DIMM_SPD_TYPE, 11 },
1105 		{ PDC_DIMM_SPD_FRESH_RATE, 12 },
1106 		{ PDC_DIMM_SPD_COLUMN_NUM, 4 },
1107 		{ PDC_DIMM_SPD_ATTRIBUTE, 21 },
1108 		{ PDC_DIMM_SPD_ROW_NUM, 3 },
1109 		{ PDC_DIMM_SPD_BANK_NUM, 17 },
1110 		{ PDC_DIMM_SPD_MODULE_ROW, 5 },
1111 		{ PDC_DIMM_SPD_ROW_PRE_CHARGE, 27 },
1112 		{ PDC_DIMM_SPD_ROW_ACTIVE_DELAY, 28 },
1113 		{ PDC_DIMM_SPD_RAS_CAS_DELAY, 29 },
1114 		{ PDC_DIMM_SPD_ACTIVE_PRECHARGE, 30 },
1115 		{ PDC_DIMM_SPD_CAS_LATENCY, 18 },
1116 	};
1117 
1118 	/* hard-code chip #0 */
1119 	mmio += PDC_CHIP0_OFS;
1120 
1121 	for (i = 0; i < ARRAY_SIZE(pdc_i2c_read_data); i++)
1122 		pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS,
1123 				  pdc_i2c_read_data[i].reg,
1124 				  &spd0[pdc_i2c_read_data[i].ofs]);
1125 
1126 	data |= (spd0[4] - 8) | ((spd0[21] != 0) << 3) | ((spd0[3]-11) << 4);
1127 	data |= ((spd0[17] / 4) << 6) | ((spd0[5] / 2) << 7) |
1128 		((((spd0[27] + 9) / 10) - 1) << 8) ;
1129 	data |= (((((spd0[29] > spd0[28])
1130 		    ? spd0[29] : spd0[28]) + 9) / 10) - 1) << 10;
1131 	data |= ((spd0[30] - spd0[29] + 9) / 10 - 2) << 12;
1132 
1133 	if (spd0[18] & 0x08)
1134 		data |= ((0x03) << 14);
1135 	else if (spd0[18] & 0x04)
1136 		data |= ((0x02) << 14);
1137 	else if (spd0[18] & 0x01)
1138 		data |= ((0x01) << 14);
1139 	else
1140 		data |= (0 << 14);
1141 
1142 	/*
1143 	   Calculate the size of bDIMMSize (power of 2) and
1144 	   merge the DIMM size by program start/end address.
1145 	*/
1146 
1147 	bdimmsize = spd0[4] + (spd0[5] / 2) + spd0[3] + (spd0[17] / 2) + 3;
1148 	size = (1 << bdimmsize) >> 20;	/* size = xxx(MB) */
1149 	data |= (((size / 16) - 1) << 16);
1150 	data |= (0 << 23);
1151 	data |= 8;
1152 	writel(data, mmio + PDC_DIMM0_CONTROL);
1153 	readl(mmio + PDC_DIMM0_CONTROL);
1154 	return size;
1155 }
1156 
1157 
1158 static unsigned int pdc20621_prog_dimm_global(struct ata_host *host)
1159 {
1160 	u32 data, spd0;
1161 	int error, i;
1162 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
1163 
1164 	/* hard-code chip #0 */
1165 	mmio += PDC_CHIP0_OFS;
1166 
1167 	/*
1168 	  Set To Default : DIMM Module Global Control Register (0x022259F1)
1169 	  DIMM Arbitration Disable (bit 20)
1170 	  DIMM Data/Control Output Driving Selection (bit12 - bit15)
1171 	  Refresh Enable (bit 17)
1172 	*/
1173 
1174 	data = 0x022259F1;
1175 	writel(data, mmio + PDC_SDRAM_CONTROL);
1176 	readl(mmio + PDC_SDRAM_CONTROL);
1177 
1178 	/* Turn on for ECC */
1179 	if (!pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS,
1180 			       PDC_DIMM_SPD_TYPE, &spd0)) {
1181 		dev_err(host->dev,
1182 			"Failed in i2c read: device=%#x, subaddr=%#x\n",
1183 			PDC_DIMM0_SPD_DEV_ADDRESS, PDC_DIMM_SPD_TYPE);
1184 		return 1;
1185 	}
1186 	if (spd0 == 0x02) {
1187 		data |= (0x01 << 16);
1188 		writel(data, mmio + PDC_SDRAM_CONTROL);
1189 		readl(mmio + PDC_SDRAM_CONTROL);
1190 		dev_err(host->dev, "Local DIMM ECC Enabled\n");
1191 	}
1192 
1193 	/* DIMM Initialization Select/Enable (bit 18/19) */
1194 	data &= (~(1<<18));
1195 	data |= (1<<19);
1196 	writel(data, mmio + PDC_SDRAM_CONTROL);
1197 
1198 	error = 1;
1199 	for (i = 1; i <= 10; i++) {   /* polling ~5 secs */
1200 		data = readl(mmio + PDC_SDRAM_CONTROL);
1201 		if (!(data & (1<<19))) {
1202 			error = 0;
1203 			break;
1204 		}
1205 		msleep(i*100);
1206 	}
1207 	return error;
1208 }
1209 
1210 
1211 static unsigned int pdc20621_dimm_init(struct ata_host *host)
1212 {
1213 	int speed, size, length;
1214 	u32 addr, spd0, pci_status;
1215 	u32 time_period = 0;
1216 	u32 tcount = 0;
1217 	u32 ticks = 0;
1218 	u32 clock = 0;
1219 	u32 fparam = 0;
1220 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
1221 
1222 	/* hard-code chip #0 */
1223 	mmio += PDC_CHIP0_OFS;
1224 
1225 	/* Initialize PLL based upon PCI Bus Frequency */
1226 
1227 	/* Initialize Time Period Register */
1228 	writel(0xffffffff, mmio + PDC_TIME_PERIOD);
1229 	time_period = readl(mmio + PDC_TIME_PERIOD);
1230 	dev_dbg(host->dev, "Time Period Register (0x40): 0x%x\n", time_period);
1231 
1232 	/* Enable timer */
1233 	writel(PDC_TIMER_DEFAULT, mmio + PDC_TIME_CONTROL);
1234 	readl(mmio + PDC_TIME_CONTROL);
1235 
1236 	/* Wait 3 seconds */
1237 	msleep(3000);
1238 
1239 	/*
1240 	   When timer is enabled, counter is decreased every internal
1241 	   clock cycle.
1242 	*/
1243 
1244 	tcount = readl(mmio + PDC_TIME_COUNTER);
1245 	dev_dbg(host->dev, "Time Counter Register (0x44): 0x%x\n", tcount);
1246 
1247 	/*
1248 	   If SX4 is on PCI-X bus, after 3 seconds, the timer counter
1249 	   register should be >= (0xffffffff - 3x10^8).
1250 	*/
1251 	if (tcount >= PCI_X_TCOUNT) {
1252 		ticks = (time_period - tcount);
1253 		dev_dbg(host->dev, "Num counters 0x%x (%d)\n", ticks, ticks);
1254 
1255 		clock = (ticks / 300000);
1256 		dev_dbg(host->dev, "10 * Internal clk = 0x%x (%d)\n",
1257 			clock, clock);
1258 
1259 		clock = (clock * 33);
1260 		dev_dbg(host->dev, "10 * Internal clk * 33 = 0x%x (%d)\n",
1261 			clock, clock);
1262 
1263 		/* PLL F Param (bit 22:16) */
1264 		fparam = (1400000 / clock) - 2;
1265 		dev_dbg(host->dev, "PLL F Param: 0x%x (%d)\n", fparam, fparam);
1266 
1267 		/* OD param = 0x2 (bit 31:30), R param = 0x5 (bit 29:25) */
1268 		pci_status = (0x8a001824 | (fparam << 16));
1269 	} else
1270 		pci_status = PCI_PLL_INIT;
1271 
1272 	/* Initialize PLL. */
1273 	dev_dbg(host->dev, "pci_status: 0x%x\n", pci_status);
1274 	writel(pci_status, mmio + PDC_CTL_STATUS);
1275 	readl(mmio + PDC_CTL_STATUS);
1276 
1277 	/*
1278 	   Read SPD of DIMM by I2C interface,
1279 	   and program the DIMM Module Controller.
1280 	*/
1281 	if (!(speed = pdc20621_detect_dimm(host))) {
1282 		dev_err(host->dev, "Detect Local DIMM Fail\n");
1283 		return 1;	/* DIMM error */
1284 	}
1285 	dev_dbg(host->dev, "Local DIMM Speed = %d\n", speed);
1286 
1287 	/* Programming DIMM0 Module Control Register (index_CID0:80h) */
1288 	size = pdc20621_prog_dimm0(host);
1289 	dev_dbg(host->dev, "Local DIMM Size = %dMB\n", size);
1290 
1291 	/* Programming DIMM Module Global Control Register (index_CID0:88h) */
1292 	if (pdc20621_prog_dimm_global(host)) {
1293 		dev_err(host->dev,
1294 			"Programming DIMM Module Global Control Register Fail\n");
1295 		return 1;
1296 	}
1297 
1298 	if (dimm_test) {
1299 		u8 test_parttern1[40] =
1300 			{0x55,0xAA,'P','r','o','m','i','s','e',' ',
1301 			'N','o','t',' ','Y','e','t',' ',
1302 			'D','e','f','i','n','e','d',' ',
1303 			'1','.','1','0',
1304 			'9','8','0','3','1','6','1','2',0,0};
1305 		u8 test_parttern2[40] = {0};
1306 
1307 		pdc20621_put_to_dimm(host, test_parttern2, 0x10040, 40);
1308 		pdc20621_put_to_dimm(host, test_parttern2, 0x40, 40);
1309 
1310 		pdc20621_put_to_dimm(host, test_parttern1, 0x10040, 40);
1311 		pdc20621_get_from_dimm(host, test_parttern2, 0x40, 40);
1312 		dev_info(host->dev, "DIMM test pattern 1: %x, %x, %s\n", test_parttern2[0],
1313 		       test_parttern2[1], &(test_parttern2[2]));
1314 		pdc20621_get_from_dimm(host, test_parttern2, 0x10040,
1315 				       40);
1316 		dev_info(host->dev, "DIMM test pattern 2: %x, %x, %s\n",
1317 			 test_parttern2[0],
1318 			 test_parttern2[1], &(test_parttern2[2]));
1319 
1320 		pdc20621_put_to_dimm(host, test_parttern1, 0x40, 40);
1321 		pdc20621_get_from_dimm(host, test_parttern2, 0x40, 40);
1322 		dev_info(host->dev, "DIMM test pattern 3: %x, %x, %s\n",
1323 			 test_parttern2[0],
1324 			 test_parttern2[1], &(test_parttern2[2]));
1325 	}
1326 
1327 	/* ECC initiliazation. */
1328 
1329 	if (!pdc20621_i2c_read(host, PDC_DIMM0_SPD_DEV_ADDRESS,
1330 			       PDC_DIMM_SPD_TYPE, &spd0)) {
1331 		dev_err(host->dev,
1332 			"Failed in i2c read: device=%#x, subaddr=%#x\n",
1333 		       PDC_DIMM0_SPD_DEV_ADDRESS, PDC_DIMM_SPD_TYPE);
1334 		return 1;
1335 	}
1336 	if (spd0 == 0x02) {
1337 		void *buf;
1338 		dev_dbg(host->dev, "Start ECC initialization\n");
1339 		addr = 0;
1340 		length = size * 1024 * 1024;
1341 		buf = kzalloc(ECC_ERASE_BUF_SZ, GFP_KERNEL);
1342 		if (!buf)
1343 			return 1;
1344 		while (addr < length) {
1345 			pdc20621_put_to_dimm(host, buf, addr,
1346 					     ECC_ERASE_BUF_SZ);
1347 			addr += ECC_ERASE_BUF_SZ;
1348 		}
1349 		kfree(buf);
1350 		dev_dbg(host->dev, "Finish ECC initialization\n");
1351 	}
1352 	return 0;
1353 }
1354 
1355 
1356 static void pdc_20621_init(struct ata_host *host)
1357 {
1358 	u32 tmp;
1359 	void __iomem *mmio = host->iomap[PDC_MMIO_BAR];
1360 
1361 	/* hard-code chip #0 */
1362 	mmio += PDC_CHIP0_OFS;
1363 
1364 	/*
1365 	 * Select page 0x40 for our 32k DIMM window
1366 	 */
1367 	tmp = readl(mmio + PDC_20621_DIMM_WINDOW) & 0xffff0000;
1368 	tmp |= PDC_PAGE_WINDOW;	/* page 40h; arbitrarily selected */
1369 	writel(tmp, mmio + PDC_20621_DIMM_WINDOW);
1370 
1371 	/*
1372 	 * Reset Host DMA
1373 	 */
1374 	tmp = readl(mmio + PDC_HDMA_CTLSTAT);
1375 	tmp |= PDC_RESET;
1376 	writel(tmp, mmio + PDC_HDMA_CTLSTAT);
1377 	readl(mmio + PDC_HDMA_CTLSTAT);		/* flush */
1378 
1379 	udelay(10);
1380 
1381 	tmp = readl(mmio + PDC_HDMA_CTLSTAT);
1382 	tmp &= ~PDC_RESET;
1383 	writel(tmp, mmio + PDC_HDMA_CTLSTAT);
1384 	readl(mmio + PDC_HDMA_CTLSTAT);		/* flush */
1385 }
1386 
1387 static int pdc_sata_init_one(struct pci_dev *pdev,
1388 			     const struct pci_device_id *ent)
1389 {
1390 	const struct ata_port_info *ppi[] =
1391 		{ &pdc_port_info[ent->driver_data], NULL };
1392 	struct ata_host *host;
1393 	struct pdc_host_priv *hpriv;
1394 	int i, rc;
1395 
1396 	ata_print_version_once(&pdev->dev, DRV_VERSION);
1397 
1398 	/* allocate host */
1399 	host = ata_host_alloc_pinfo(&pdev->dev, ppi, 4);
1400 	hpriv = devm_kzalloc(&pdev->dev, sizeof(*hpriv), GFP_KERNEL);
1401 	if (!host || !hpriv)
1402 		return -ENOMEM;
1403 
1404 	host->private_data = hpriv;
1405 
1406 	/* acquire resources and fill host */
1407 	rc = pcim_enable_device(pdev);
1408 	if (rc)
1409 		return rc;
1410 
1411 	rc = pcim_iomap_regions(pdev, (1 << PDC_MMIO_BAR) | (1 << PDC_DIMM_BAR),
1412 				DRV_NAME);
1413 	if (rc == -EBUSY)
1414 		pcim_pin_device(pdev);
1415 	if (rc)
1416 		return rc;
1417 	host->iomap = pcim_iomap_table(pdev);
1418 
1419 	for (i = 0; i < 4; i++) {
1420 		struct ata_port *ap = host->ports[i];
1421 		void __iomem *base = host->iomap[PDC_MMIO_BAR] + PDC_CHIP0_OFS;
1422 		unsigned int offset = 0x200 + i * 0x80;
1423 
1424 		pdc_sata_setup_port(&ap->ioaddr, base + offset);
1425 
1426 		ata_port_pbar_desc(ap, PDC_MMIO_BAR, -1, "mmio");
1427 		ata_port_pbar_desc(ap, PDC_DIMM_BAR, -1, "dimm");
1428 		ata_port_pbar_desc(ap, PDC_MMIO_BAR, offset, "port");
1429 	}
1430 
1431 	/* configure and activate */
1432 	rc = dma_set_mask_and_coherent(&pdev->dev, ATA_DMA_MASK);
1433 	if (rc)
1434 		return rc;
1435 
1436 	if (pdc20621_dimm_init(host))
1437 		return -ENOMEM;
1438 	pdc_20621_init(host);
1439 
1440 	pci_set_master(pdev);
1441 	return ata_host_activate(host, pdev->irq, pdc20621_interrupt,
1442 				 IRQF_SHARED, &pdc_sata_sht);
1443 }
1444 
1445 module_pci_driver(pdc_sata_pci_driver);
1446 
1447 MODULE_AUTHOR("Jeff Garzik");
1448 MODULE_DESCRIPTION("Promise SATA low-level driver");
1449 MODULE_LICENSE("GPL");
1450 MODULE_DEVICE_TABLE(pci, pdc_sata_pci_tbl);
1451 MODULE_VERSION(DRV_VERSION);
1452