xref: /openbmc/linux/drivers/scsi/mesh.c (revision c21b37f6)
1 /*
2  * SCSI low-level driver for the MESH (Macintosh Enhanced SCSI Hardware)
3  * bus adaptor found on Power Macintosh computers.
4  * We assume the MESH is connected to a DBDMA (descriptor-based DMA)
5  * controller.
6  *
7  * Paul Mackerras, August 1996.
8  * Copyright (C) 1996 Paul Mackerras.
9  *
10  * Apr. 21 2002  - BenH		Rework bus reset code for new error handler
11  *                              Add delay after initial bus reset
12  *                              Add module parameters
13  *
14  * Sep. 27 2003  - BenH		Move to new driver model, fix some write posting
15  *				issues
16  * To do:
17  * - handle aborts correctly
18  * - retry arbitration if lost (unless higher levels do this for us)
19  * - power down the chip when no device is detected
20  */
21 #include <linux/module.h>
22 #include <linux/kernel.h>
23 #include <linux/delay.h>
24 #include <linux/types.h>
25 #include <linux/string.h>
26 #include <linux/slab.h>
27 #include <linux/blkdev.h>
28 #include <linux/proc_fs.h>
29 #include <linux/stat.h>
30 #include <linux/interrupt.h>
31 #include <linux/reboot.h>
32 #include <linux/spinlock.h>
33 #include <asm/dbdma.h>
34 #include <asm/io.h>
35 #include <asm/pgtable.h>
36 #include <asm/prom.h>
37 #include <asm/system.h>
38 #include <asm/irq.h>
39 #include <asm/hydra.h>
40 #include <asm/processor.h>
41 #include <asm/machdep.h>
42 #include <asm/pmac_feature.h>
43 #include <asm/pci-bridge.h>
44 #include <asm/macio.h>
45 
46 #include <scsi/scsi.h>
47 #include <scsi/scsi_cmnd.h>
48 #include <scsi/scsi_device.h>
49 #include <scsi/scsi_host.h>
50 
51 #include "mesh.h"
52 
53 #if 1
54 #undef KERN_DEBUG
55 #define KERN_DEBUG KERN_WARNING
56 #endif
57 
58 MODULE_AUTHOR("Paul Mackerras (paulus@samba.org)");
59 MODULE_DESCRIPTION("PowerMac MESH SCSI driver");
60 MODULE_LICENSE("GPL");
61 
62 static int sync_rate = CONFIG_SCSI_MESH_SYNC_RATE;
63 static int sync_targets = 0xff;
64 static int resel_targets = 0xff;
65 static int debug_targets = 0;	/* print debug for these targets */
66 static int init_reset_delay = CONFIG_SCSI_MESH_RESET_DELAY_MS;
67 
68 module_param(sync_rate, int, 0);
69 MODULE_PARM_DESC(sync_rate, "Synchronous rate (0..10, 0=async)");
70 module_param(sync_targets, int, 0);
71 MODULE_PARM_DESC(sync_targets, "Bitmask of targets allowed to set synchronous");
72 module_param(resel_targets, int, 0);
73 MODULE_PARM_DESC(resel_targets, "Bitmask of targets allowed to set disconnect");
74 module_param(debug_targets, int, 0644);
75 MODULE_PARM_DESC(debug_targets, "Bitmask of debugged targets");
76 module_param(init_reset_delay, int, 0);
77 MODULE_PARM_DESC(init_reset_delay, "Initial bus reset delay (0=no reset)");
78 
79 static int mesh_sync_period = 100;
80 static int mesh_sync_offset = 0;
81 static unsigned char use_active_neg = 0;  /* bit mask for SEQ_ACTIVE_NEG if used */
82 
83 #define ALLOW_SYNC(tgt)		((sync_targets >> (tgt)) & 1)
84 #define ALLOW_RESEL(tgt)	((resel_targets >> (tgt)) & 1)
85 #define ALLOW_DEBUG(tgt)	((debug_targets >> (tgt)) & 1)
86 #define DEBUG_TARGET(cmd)	((cmd) && ALLOW_DEBUG((cmd)->device->id))
87 
88 #undef MESH_DBG
89 #define N_DBG_LOG	50
90 #define N_DBG_SLOG	20
91 #define NUM_DBG_EVENTS	13
92 #undef	DBG_USE_TB		/* bombs on 601 */
93 
94 struct dbglog {
95 	char	*fmt;
96 	u32	tb;
97 	u8	phase;
98 	u8	bs0;
99 	u8	bs1;
100 	u8	tgt;
101 	int	d;
102 };
103 
104 enum mesh_phase {
105 	idle,
106 	arbitrating,
107 	selecting,
108 	commanding,
109 	dataing,
110 	statusing,
111 	busfreeing,
112 	disconnecting,
113 	reselecting,
114 	sleeping
115 };
116 
117 enum msg_phase {
118 	msg_none,
119 	msg_out,
120 	msg_out_xxx,
121 	msg_out_last,
122 	msg_in,
123 	msg_in_bad,
124 };
125 
126 enum sdtr_phase {
127 	do_sdtr,
128 	sdtr_sent,
129 	sdtr_done
130 };
131 
132 struct mesh_target {
133 	enum sdtr_phase sdtr_state;
134 	int	sync_params;
135 	int	data_goes_out;		/* guess as to data direction */
136 	struct scsi_cmnd *current_req;
137 	u32	saved_ptr;
138 #ifdef MESH_DBG
139 	int	log_ix;
140 	int	n_log;
141 	struct dbglog log[N_DBG_LOG];
142 #endif
143 };
144 
145 struct mesh_state {
146 	volatile struct	mesh_regs __iomem *mesh;
147 	int	meshintr;
148 	volatile struct	dbdma_regs __iomem *dma;
149 	int	dmaintr;
150 	struct	Scsi_Host *host;
151 	struct	mesh_state *next;
152 	struct scsi_cmnd *request_q;
153 	struct scsi_cmnd *request_qtail;
154 	enum mesh_phase phase;		/* what we're currently trying to do */
155 	enum msg_phase msgphase;
156 	int	conn_tgt;		/* target we're connected to */
157 	struct scsi_cmnd *current_req;		/* req we're currently working on */
158 	int	data_ptr;
159 	int	dma_started;
160 	int	dma_count;
161 	int	stat;
162 	int	aborting;
163 	int	expect_reply;
164 	int	n_msgin;
165 	u8	msgin[16];
166 	int	n_msgout;
167 	int	last_n_msgout;
168 	u8	msgout[16];
169 	struct dbdma_cmd *dma_cmds;	/* space for dbdma commands, aligned */
170 	dma_addr_t dma_cmd_bus;
171 	void	*dma_cmd_space;
172 	int	dma_cmd_size;
173 	int	clk_freq;
174 	struct mesh_target tgts[8];
175 	struct macio_dev *mdev;
176 	struct pci_dev* pdev;
177 #ifdef MESH_DBG
178 	int	log_ix;
179 	int	n_log;
180 	struct dbglog log[N_DBG_SLOG];
181 #endif
182 };
183 
184 /*
185  * Driver is too messy, we need a few prototypes...
186  */
187 static void mesh_done(struct mesh_state *ms, int start_next);
188 static void mesh_interrupt(struct mesh_state *ms);
189 static void cmd_complete(struct mesh_state *ms);
190 static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd);
191 static void halt_dma(struct mesh_state *ms);
192 static void phase_mismatch(struct mesh_state *ms);
193 
194 
195 /*
196  * Some debugging & logging routines
197  */
198 
199 #ifdef MESH_DBG
200 
201 static inline u32 readtb(void)
202 {
203 	u32 tb;
204 
205 #ifdef DBG_USE_TB
206 	/* Beware: if you enable this, it will crash on 601s. */
207 	asm ("mftb %0" : "=r" (tb) : );
208 #else
209 	tb = 0;
210 #endif
211 	return tb;
212 }
213 
214 static void dlog(struct mesh_state *ms, char *fmt, int a)
215 {
216 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
217 	struct dbglog *tlp, *slp;
218 
219 	tlp = &tp->log[tp->log_ix];
220 	slp = &ms->log[ms->log_ix];
221 	tlp->fmt = fmt;
222 	tlp->tb = readtb();
223 	tlp->phase = (ms->msgphase << 4) + ms->phase;
224 	tlp->bs0 = ms->mesh->bus_status0;
225 	tlp->bs1 = ms->mesh->bus_status1;
226 	tlp->tgt = ms->conn_tgt;
227 	tlp->d = a;
228 	*slp = *tlp;
229 	if (++tp->log_ix >= N_DBG_LOG)
230 		tp->log_ix = 0;
231 	if (tp->n_log < N_DBG_LOG)
232 		++tp->n_log;
233 	if (++ms->log_ix >= N_DBG_SLOG)
234 		ms->log_ix = 0;
235 	if (ms->n_log < N_DBG_SLOG)
236 		++ms->n_log;
237 }
238 
239 static void dumplog(struct mesh_state *ms, int t)
240 {
241 	struct mesh_target *tp = &ms->tgts[t];
242 	struct dbglog *lp;
243 	int i;
244 
245 	if (tp->n_log == 0)
246 		return;
247 	i = tp->log_ix - tp->n_log;
248 	if (i < 0)
249 		i += N_DBG_LOG;
250 	tp->n_log = 0;
251 	do {
252 		lp = &tp->log[i];
253 		printk(KERN_DEBUG "mesh log %d: bs=%.2x%.2x ph=%.2x ",
254 		       t, lp->bs1, lp->bs0, lp->phase);
255 #ifdef DBG_USE_TB
256 		printk("tb=%10u ", lp->tb);
257 #endif
258 		printk(lp->fmt, lp->d);
259 		printk("\n");
260 		if (++i >= N_DBG_LOG)
261 			i = 0;
262 	} while (i != tp->log_ix);
263 }
264 
265 static void dumpslog(struct mesh_state *ms)
266 {
267 	struct dbglog *lp;
268 	int i;
269 
270 	if (ms->n_log == 0)
271 		return;
272 	i = ms->log_ix - ms->n_log;
273 	if (i < 0)
274 		i += N_DBG_SLOG;
275 	ms->n_log = 0;
276 	do {
277 		lp = &ms->log[i];
278 		printk(KERN_DEBUG "mesh log: bs=%.2x%.2x ph=%.2x t%d ",
279 		       lp->bs1, lp->bs0, lp->phase, lp->tgt);
280 #ifdef DBG_USE_TB
281 		printk("tb=%10u ", lp->tb);
282 #endif
283 		printk(lp->fmt, lp->d);
284 		printk("\n");
285 		if (++i >= N_DBG_SLOG)
286 			i = 0;
287 	} while (i != ms->log_ix);
288 }
289 
290 #else
291 
292 static inline void dlog(struct mesh_state *ms, char *fmt, int a)
293 {}
294 static inline void dumplog(struct mesh_state *ms, int tgt)
295 {}
296 static inline void dumpslog(struct mesh_state *ms)
297 {}
298 
299 #endif /* MESH_DBG */
300 
301 #define MKWORD(a, b, c, d)	(((a) << 24) + ((b) << 16) + ((c) << 8) + (d))
302 
303 static void
304 mesh_dump_regs(struct mesh_state *ms)
305 {
306 	volatile struct mesh_regs __iomem *mr = ms->mesh;
307 	volatile struct dbdma_regs __iomem *md = ms->dma;
308 	int t;
309 	struct mesh_target *tp;
310 
311 	printk(KERN_DEBUG "mesh: state at %p, regs at %p, dma at %p\n",
312 	       ms, mr, md);
313 	printk(KERN_DEBUG "    ct=%4x seq=%2x bs=%4x fc=%2x "
314 	       "exc=%2x err=%2x im=%2x int=%2x sp=%2x\n",
315 	       (mr->count_hi << 8) + mr->count_lo, mr->sequence,
316 	       (mr->bus_status1 << 8) + mr->bus_status0, mr->fifo_count,
317 	       mr->exception, mr->error, mr->intr_mask, mr->interrupt,
318 	       mr->sync_params);
319 	while(in_8(&mr->fifo_count))
320 		printk(KERN_DEBUG " fifo data=%.2x\n",in_8(&mr->fifo));
321 	printk(KERN_DEBUG "    dma stat=%x cmdptr=%x\n",
322 	       in_le32(&md->status), in_le32(&md->cmdptr));
323 	printk(KERN_DEBUG "    phase=%d msgphase=%d conn_tgt=%d data_ptr=%d\n",
324 	       ms->phase, ms->msgphase, ms->conn_tgt, ms->data_ptr);
325 	printk(KERN_DEBUG "    dma_st=%d dma_ct=%d n_msgout=%d\n",
326 	       ms->dma_started, ms->dma_count, ms->n_msgout);
327 	for (t = 0; t < 8; ++t) {
328 		tp = &ms->tgts[t];
329 		if (tp->current_req == NULL)
330 			continue;
331 		printk(KERN_DEBUG "    target %d: req=%p goes_out=%d saved_ptr=%d\n",
332 		       t, tp->current_req, tp->data_goes_out, tp->saved_ptr);
333 	}
334 }
335 
336 
337 /*
338  * Flush write buffers on the bus path to the mesh
339  */
340 static inline void mesh_flush_io(volatile struct mesh_regs __iomem *mr)
341 {
342 	(void)in_8(&mr->mesh_id);
343 }
344 
345 
346 /*
347  * Complete a SCSI command
348  */
349 static void mesh_completed(struct mesh_state *ms, struct scsi_cmnd *cmd)
350 {
351 	(*cmd->scsi_done)(cmd);
352 }
353 
354 
355 /* Called with  meshinterrupt disabled, initialize the chipset
356  * and eventually do the initial bus reset. The lock must not be
357  * held since we can schedule.
358  */
359 static void mesh_init(struct mesh_state *ms)
360 {
361 	volatile struct mesh_regs __iomem *mr = ms->mesh;
362 	volatile struct dbdma_regs __iomem *md = ms->dma;
363 
364 	mesh_flush_io(mr);
365 	udelay(100);
366 
367 	/* Reset controller */
368 	out_le32(&md->control, (RUN|PAUSE|FLUSH|WAKE) << 16);	/* stop dma */
369 	out_8(&mr->exception, 0xff);	/* clear all exception bits */
370 	out_8(&mr->error, 0xff);	/* clear all error bits */
371 	out_8(&mr->sequence, SEQ_RESETMESH);
372 	mesh_flush_io(mr);
373 	udelay(10);
374 	out_8(&mr->intr_mask, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
375 	out_8(&mr->source_id, ms->host->this_id);
376 	out_8(&mr->sel_timeout, 25);	/* 250ms */
377 	out_8(&mr->sync_params, ASYNC_PARAMS);
378 
379 	if (init_reset_delay) {
380 		printk(KERN_INFO "mesh: performing initial bus reset...\n");
381 
382 		/* Reset bus */
383 		out_8(&mr->bus_status1, BS1_RST);	/* assert RST */
384 		mesh_flush_io(mr);
385 		udelay(30);			/* leave it on for >= 25us */
386 		out_8(&mr->bus_status1, 0);	/* negate RST */
387 		mesh_flush_io(mr);
388 
389 		/* Wait for bus to come back */
390 		msleep(init_reset_delay);
391 	}
392 
393 	/* Reconfigure controller */
394 	out_8(&mr->interrupt, 0xff);	/* clear all interrupt bits */
395 	out_8(&mr->sequence, SEQ_FLUSHFIFO);
396 	mesh_flush_io(mr);
397 	udelay(1);
398 	out_8(&mr->sync_params, ASYNC_PARAMS);
399 	out_8(&mr->sequence, SEQ_ENBRESEL);
400 
401 	ms->phase = idle;
402 	ms->msgphase = msg_none;
403 }
404 
405 
406 static void mesh_start_cmd(struct mesh_state *ms, struct scsi_cmnd *cmd)
407 {
408 	volatile struct mesh_regs __iomem *mr = ms->mesh;
409 	int t, id;
410 
411 	id = cmd->device->id;
412 	ms->current_req = cmd;
413 	ms->tgts[id].data_goes_out = cmd->sc_data_direction == DMA_TO_DEVICE;
414 	ms->tgts[id].current_req = cmd;
415 
416 #if 1
417 	if (DEBUG_TARGET(cmd)) {
418 		int i;
419 		printk(KERN_DEBUG "mesh_start: %p ser=%lu tgt=%d cmd=",
420 		       cmd, cmd->serial_number, id);
421 		for (i = 0; i < cmd->cmd_len; ++i)
422 			printk(" %x", cmd->cmnd[i]);
423 		printk(" use_sg=%d buffer=%p bufflen=%u\n",
424 		       scsi_sg_count(cmd), scsi_sglist(cmd), scsi_bufflen(cmd));
425 	}
426 #endif
427 	if (ms->dma_started)
428 		panic("mesh: double DMA start !\n");
429 
430 	ms->phase = arbitrating;
431 	ms->msgphase = msg_none;
432 	ms->data_ptr = 0;
433 	ms->dma_started = 0;
434 	ms->n_msgout = 0;
435 	ms->last_n_msgout = 0;
436 	ms->expect_reply = 0;
437 	ms->conn_tgt = id;
438 	ms->tgts[id].saved_ptr = 0;
439 	ms->stat = DID_OK;
440 	ms->aborting = 0;
441 #ifdef MESH_DBG
442 	ms->tgts[id].n_log = 0;
443 	dlog(ms, "start cmd=%x", (int) cmd);
444 #endif
445 
446 	/* Off we go */
447 	dlog(ms, "about to arb, intr/exc/err/fc=%.8x",
448 	     MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
449 	out_8(&mr->interrupt, INT_CMDDONE);
450 	out_8(&mr->sequence, SEQ_ENBRESEL);
451 	mesh_flush_io(mr);
452 	udelay(1);
453 
454 	if (in_8(&mr->bus_status1) & (BS1_BSY | BS1_SEL)) {
455 		/*
456 		 * Some other device has the bus or is arbitrating for it -
457 		 * probably a target which is about to reselect us.
458 		 */
459 		dlog(ms, "busy b4 arb, intr/exc/err/fc=%.8x",
460 		     MKWORD(mr->interrupt, mr->exception,
461 			    mr->error, mr->fifo_count));
462 		for (t = 100; t > 0; --t) {
463 			if ((in_8(&mr->bus_status1) & (BS1_BSY | BS1_SEL)) == 0)
464 				break;
465 			if (in_8(&mr->interrupt) != 0) {
466 				dlog(ms, "intr b4 arb, intr/exc/err/fc=%.8x",
467 				     MKWORD(mr->interrupt, mr->exception,
468 					    mr->error, mr->fifo_count));
469 				mesh_interrupt(ms);
470 				if (ms->phase != arbitrating)
471 					return;
472 			}
473 			udelay(1);
474 		}
475 		if (in_8(&mr->bus_status1) & (BS1_BSY | BS1_SEL)) {
476 			/* XXX should try again in a little while */
477 			ms->stat = DID_BUS_BUSY;
478 			ms->phase = idle;
479 			mesh_done(ms, 0);
480 			return;
481 		}
482 	}
483 
484 	/*
485 	 * Apparently the mesh has a bug where it will assert both its
486 	 * own bit and the target's bit on the bus during arbitration.
487 	 */
488 	out_8(&mr->dest_id, mr->source_id);
489 
490 	/*
491 	 * There appears to be a race with reselection sometimes,
492 	 * where a target reselects us just as we issue the
493 	 * arbitrate command.  It seems that then the arbitrate
494 	 * command just hangs waiting for the bus to be free
495 	 * without giving us a reselection exception.
496 	 * The only way I have found to get it to respond correctly
497 	 * is this: disable reselection before issuing the arbitrate
498 	 * command, then after issuing it, if it looks like a target
499 	 * is trying to reselect us, reset the mesh and then enable
500 	 * reselection.
501 	 */
502 	out_8(&mr->sequence, SEQ_DISRESEL);
503 	if (in_8(&mr->interrupt) != 0) {
504 		dlog(ms, "intr after disresel, intr/exc/err/fc=%.8x",
505 		     MKWORD(mr->interrupt, mr->exception,
506 			    mr->error, mr->fifo_count));
507 		mesh_interrupt(ms);
508 		if (ms->phase != arbitrating)
509 			return;
510 		dlog(ms, "after intr after disresel, intr/exc/err/fc=%.8x",
511 		     MKWORD(mr->interrupt, mr->exception,
512 			    mr->error, mr->fifo_count));
513 	}
514 
515 	out_8(&mr->sequence, SEQ_ARBITRATE);
516 
517 	for (t = 230; t > 0; --t) {
518 		if (in_8(&mr->interrupt) != 0)
519 			break;
520 		udelay(1);
521 	}
522 	dlog(ms, "after arb, intr/exc/err/fc=%.8x",
523 	     MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
524 	if (in_8(&mr->interrupt) == 0 && (in_8(&mr->bus_status1) & BS1_SEL)
525 	    && (in_8(&mr->bus_status0) & BS0_IO)) {
526 		/* looks like a reselection - try resetting the mesh */
527 		dlog(ms, "resel? after arb, intr/exc/err/fc=%.8x",
528 		     MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
529 		out_8(&mr->sequence, SEQ_RESETMESH);
530 		mesh_flush_io(mr);
531 		udelay(10);
532 		out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
533 		out_8(&mr->intr_mask, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
534 		out_8(&mr->sequence, SEQ_ENBRESEL);
535 		mesh_flush_io(mr);
536 		for (t = 10; t > 0 && in_8(&mr->interrupt) == 0; --t)
537 			udelay(1);
538 		dlog(ms, "tried reset after arb, intr/exc/err/fc=%.8x",
539 		     MKWORD(mr->interrupt, mr->exception, mr->error, mr->fifo_count));
540 #ifndef MESH_MULTIPLE_HOSTS
541 		if (in_8(&mr->interrupt) == 0 && (in_8(&mr->bus_status1) & BS1_SEL)
542 		    && (in_8(&mr->bus_status0) & BS0_IO)) {
543 			printk(KERN_ERR "mesh: controller not responding"
544 			       " to reselection!\n");
545 			/*
546 			 * If this is a target reselecting us, and the
547 			 * mesh isn't responding, the higher levels of
548 			 * the scsi code will eventually time out and
549 			 * reset the bus.
550 			 */
551 		}
552 #endif
553 	}
554 }
555 
556 /*
557  * Start the next command for a MESH.
558  * Should be called with interrupts disabled.
559  */
560 static void mesh_start(struct mesh_state *ms)
561 {
562 	struct scsi_cmnd *cmd, *prev, *next;
563 
564 	if (ms->phase != idle || ms->current_req != NULL) {
565 		printk(KERN_ERR "inappropriate mesh_start (phase=%d, ms=%p)",
566 		       ms->phase, ms);
567 		return;
568 	}
569 
570 	while (ms->phase == idle) {
571 		prev = NULL;
572 		for (cmd = ms->request_q; ; cmd = (struct scsi_cmnd *) cmd->host_scribble) {
573 			if (cmd == NULL)
574 				return;
575 			if (ms->tgts[cmd->device->id].current_req == NULL)
576 				break;
577 			prev = cmd;
578 		}
579 		next = (struct scsi_cmnd *) cmd->host_scribble;
580 		if (prev == NULL)
581 			ms->request_q = next;
582 		else
583 			prev->host_scribble = (void *) next;
584 		if (next == NULL)
585 			ms->request_qtail = prev;
586 
587 		mesh_start_cmd(ms, cmd);
588 	}
589 }
590 
591 static void mesh_done(struct mesh_state *ms, int start_next)
592 {
593 	struct scsi_cmnd *cmd;
594 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
595 
596 	cmd = ms->current_req;
597 	ms->current_req = NULL;
598 	tp->current_req = NULL;
599 	if (cmd) {
600 		cmd->result = (ms->stat << 16) + cmd->SCp.Status;
601 		if (ms->stat == DID_OK)
602 			cmd->result += (cmd->SCp.Message << 8);
603 		if (DEBUG_TARGET(cmd)) {
604 			printk(KERN_DEBUG "mesh_done: result = %x, data_ptr=%d, buflen=%d\n",
605 			       cmd->result, ms->data_ptr, scsi_bufflen(cmd));
606 #if 0
607 			/* needs to use sg? */
608 			if ((cmd->cmnd[0] == 0 || cmd->cmnd[0] == 0x12 || cmd->cmnd[0] == 3)
609 			    && cmd->request_buffer != 0) {
610 				unsigned char *b = cmd->request_buffer;
611 				printk(KERN_DEBUG "buffer = %x %x %x %x %x %x %x %x\n",
612 				       b[0], b[1], b[2], b[3], b[4], b[5], b[6], b[7]);
613 			}
614 #endif
615 		}
616 		cmd->SCp.this_residual -= ms->data_ptr;
617 		mesh_completed(ms, cmd);
618 	}
619 	if (start_next) {
620 		out_8(&ms->mesh->sequence, SEQ_ENBRESEL);
621 		mesh_flush_io(ms->mesh);
622 		udelay(1);
623 		ms->phase = idle;
624 		mesh_start(ms);
625 	}
626 }
627 
628 static inline void add_sdtr_msg(struct mesh_state *ms)
629 {
630 	int i = ms->n_msgout;
631 
632 	ms->msgout[i] = EXTENDED_MESSAGE;
633 	ms->msgout[i+1] = 3;
634 	ms->msgout[i+2] = EXTENDED_SDTR;
635 	ms->msgout[i+3] = mesh_sync_period/4;
636 	ms->msgout[i+4] = (ALLOW_SYNC(ms->conn_tgt)? mesh_sync_offset: 0);
637 	ms->n_msgout = i + 5;
638 }
639 
640 static void set_sdtr(struct mesh_state *ms, int period, int offset)
641 {
642 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
643 	volatile struct mesh_regs __iomem *mr = ms->mesh;
644 	int v, tr;
645 
646 	tp->sdtr_state = sdtr_done;
647 	if (offset == 0) {
648 		/* asynchronous */
649 		if (SYNC_OFF(tp->sync_params))
650 			printk(KERN_INFO "mesh: target %d now asynchronous\n",
651 			       ms->conn_tgt);
652 		tp->sync_params = ASYNC_PARAMS;
653 		out_8(&mr->sync_params, ASYNC_PARAMS);
654 		return;
655 	}
656 	/*
657 	 * We need to compute ceil(clk_freq * period / 500e6) - 2
658 	 * without incurring overflow.
659 	 */
660 	v = (ms->clk_freq / 5000) * period;
661 	if (v <= 250000) {
662 		/* special case: sync_period == 5 * clk_period */
663 		v = 0;
664 		/* units of tr are 100kB/s */
665 		tr = (ms->clk_freq + 250000) / 500000;
666 	} else {
667 		/* sync_period == (v + 2) * 2 * clk_period */
668 		v = (v + 99999) / 100000 - 2;
669 		if (v > 15)
670 			v = 15;	/* oops */
671 		tr = ((ms->clk_freq / (v + 2)) + 199999) / 200000;
672 	}
673 	if (offset > 15)
674 		offset = 15;	/* can't happen */
675 	tp->sync_params = SYNC_PARAMS(offset, v);
676 	out_8(&mr->sync_params, tp->sync_params);
677 	printk(KERN_INFO "mesh: target %d synchronous at %d.%d MB/s\n",
678 	       ms->conn_tgt, tr/10, tr%10);
679 }
680 
681 static void start_phase(struct mesh_state *ms)
682 {
683 	int i, seq, nb;
684 	volatile struct mesh_regs __iomem *mr = ms->mesh;
685 	volatile struct dbdma_regs __iomem *md = ms->dma;
686 	struct scsi_cmnd *cmd = ms->current_req;
687 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
688 
689 	dlog(ms, "start_phase nmo/exc/fc/seq = %.8x",
690 	     MKWORD(ms->n_msgout, mr->exception, mr->fifo_count, mr->sequence));
691 	out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
692 	seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0);
693 	switch (ms->msgphase) {
694 	case msg_none:
695 		break;
696 
697 	case msg_in:
698 		out_8(&mr->count_hi, 0);
699 		out_8(&mr->count_lo, 1);
700 		out_8(&mr->sequence, SEQ_MSGIN + seq);
701 		ms->n_msgin = 0;
702 		return;
703 
704 	case msg_out:
705 		/*
706 		 * To make sure ATN drops before we assert ACK for
707 		 * the last byte of the message, we have to do the
708 		 * last byte specially.
709 		 */
710 		if (ms->n_msgout <= 0) {
711 			printk(KERN_ERR "mesh: msg_out but n_msgout=%d\n",
712 			       ms->n_msgout);
713 			mesh_dump_regs(ms);
714 			ms->msgphase = msg_none;
715 			break;
716 		}
717 		if (ALLOW_DEBUG(ms->conn_tgt)) {
718 			printk(KERN_DEBUG "mesh: sending %d msg bytes:",
719 			       ms->n_msgout);
720 			for (i = 0; i < ms->n_msgout; ++i)
721 				printk(" %x", ms->msgout[i]);
722 			printk("\n");
723 		}
724 		dlog(ms, "msgout msg=%.8x", MKWORD(ms->n_msgout, ms->msgout[0],
725 						ms->msgout[1], ms->msgout[2]));
726 		out_8(&mr->count_hi, 0);
727 		out_8(&mr->sequence, SEQ_FLUSHFIFO);
728 		mesh_flush_io(mr);
729 		udelay(1);
730 		/*
731 		 * If ATN is not already asserted, we assert it, then
732 		 * issue a SEQ_MSGOUT to get the mesh to drop ACK.
733 		 */
734 		if ((in_8(&mr->bus_status0) & BS0_ATN) == 0) {
735 			dlog(ms, "bus0 was %.2x explicitly asserting ATN", mr->bus_status0);
736 			out_8(&mr->bus_status0, BS0_ATN); /* explicit ATN */
737 			mesh_flush_io(mr);
738 			udelay(1);
739 			out_8(&mr->count_lo, 1);
740 			out_8(&mr->sequence, SEQ_MSGOUT + seq);
741 			out_8(&mr->bus_status0, 0); /* release explicit ATN */
742 			dlog(ms,"hace: after explicit ATN bus0=%.2x",mr->bus_status0);
743 		}
744 		if (ms->n_msgout == 1) {
745 			/*
746 			 * We can't issue the SEQ_MSGOUT without ATN
747 			 * until the target has asserted REQ.  The logic
748 			 * in cmd_complete handles both situations:
749 			 * REQ already asserted or not.
750 			 */
751 			cmd_complete(ms);
752 		} else {
753 			out_8(&mr->count_lo, ms->n_msgout - 1);
754 			out_8(&mr->sequence, SEQ_MSGOUT + seq);
755 			for (i = 0; i < ms->n_msgout - 1; ++i)
756 				out_8(&mr->fifo, ms->msgout[i]);
757 		}
758 		return;
759 
760 	default:
761 		printk(KERN_ERR "mesh bug: start_phase msgphase=%d\n",
762 		       ms->msgphase);
763 	}
764 
765 	switch (ms->phase) {
766 	case selecting:
767 		out_8(&mr->dest_id, ms->conn_tgt);
768 		out_8(&mr->sequence, SEQ_SELECT + SEQ_ATN);
769 		break;
770 	case commanding:
771 		out_8(&mr->sync_params, tp->sync_params);
772 		out_8(&mr->count_hi, 0);
773 		if (cmd) {
774 			out_8(&mr->count_lo, cmd->cmd_len);
775 			out_8(&mr->sequence, SEQ_COMMAND + seq);
776 			for (i = 0; i < cmd->cmd_len; ++i)
777 				out_8(&mr->fifo, cmd->cmnd[i]);
778 		} else {
779 			out_8(&mr->count_lo, 6);
780 			out_8(&mr->sequence, SEQ_COMMAND + seq);
781 			for (i = 0; i < 6; ++i)
782 				out_8(&mr->fifo, 0);
783 		}
784 		break;
785 	case dataing:
786 		/* transfer data, if any */
787 		if (!ms->dma_started) {
788 			set_dma_cmds(ms, cmd);
789 			out_le32(&md->cmdptr, virt_to_phys(ms->dma_cmds));
790 			out_le32(&md->control, (RUN << 16) | RUN);
791 			ms->dma_started = 1;
792 		}
793 		nb = ms->dma_count;
794 		if (nb > 0xfff0)
795 			nb = 0xfff0;
796 		ms->dma_count -= nb;
797 		ms->data_ptr += nb;
798 		out_8(&mr->count_lo, nb);
799 		out_8(&mr->count_hi, nb >> 8);
800 		out_8(&mr->sequence, (tp->data_goes_out?
801 				SEQ_DATAOUT: SEQ_DATAIN) + SEQ_DMA_MODE + seq);
802 		break;
803 	case statusing:
804 		out_8(&mr->count_hi, 0);
805 		out_8(&mr->count_lo, 1);
806 		out_8(&mr->sequence, SEQ_STATUS + seq);
807 		break;
808 	case busfreeing:
809 	case disconnecting:
810 		out_8(&mr->sequence, SEQ_ENBRESEL);
811 		mesh_flush_io(mr);
812 		udelay(1);
813 		dlog(ms, "enbresel intr/exc/err/fc=%.8x",
814 		     MKWORD(mr->interrupt, mr->exception, mr->error,
815 			    mr->fifo_count));
816 		out_8(&mr->sequence, SEQ_BUSFREE);
817 		break;
818 	default:
819 		printk(KERN_ERR "mesh: start_phase called with phase=%d\n",
820 		       ms->phase);
821 		dumpslog(ms);
822 	}
823 
824 }
825 
826 static inline void get_msgin(struct mesh_state *ms)
827 {
828 	volatile struct mesh_regs __iomem *mr = ms->mesh;
829 	int i, n;
830 
831 	n = mr->fifo_count;
832 	if (n != 0) {
833 		i = ms->n_msgin;
834 		ms->n_msgin = i + n;
835 		for (; n > 0; --n)
836 			ms->msgin[i++] = in_8(&mr->fifo);
837 	}
838 }
839 
840 static inline int msgin_length(struct mesh_state *ms)
841 {
842 	int b, n;
843 
844 	n = 1;
845 	if (ms->n_msgin > 0) {
846 		b = ms->msgin[0];
847 		if (b == 1) {
848 			/* extended message */
849 			n = ms->n_msgin < 2? 2: ms->msgin[1] + 2;
850 		} else if (0x20 <= b && b <= 0x2f) {
851 			/* 2-byte message */
852 			n = 2;
853 		}
854 	}
855 	return n;
856 }
857 
858 static void reselected(struct mesh_state *ms)
859 {
860 	volatile struct mesh_regs __iomem *mr = ms->mesh;
861 	struct scsi_cmnd *cmd;
862 	struct mesh_target *tp;
863 	int b, t, prev;
864 
865 	switch (ms->phase) {
866 	case idle:
867 		break;
868 	case arbitrating:
869 		if ((cmd = ms->current_req) != NULL) {
870 			/* put the command back on the queue */
871 			cmd->host_scribble = (void *) ms->request_q;
872 			if (ms->request_q == NULL)
873 				ms->request_qtail = cmd;
874 			ms->request_q = cmd;
875 			tp = &ms->tgts[cmd->device->id];
876 			tp->current_req = NULL;
877 		}
878 		break;
879 	case busfreeing:
880 		ms->phase = reselecting;
881 		mesh_done(ms, 0);
882 		break;
883 	case disconnecting:
884 		break;
885 	default:
886 		printk(KERN_ERR "mesh: reselected in phase %d/%d tgt %d\n",
887 		       ms->msgphase, ms->phase, ms->conn_tgt);
888 		dumplog(ms, ms->conn_tgt);
889 		dumpslog(ms);
890 	}
891 
892 	if (ms->dma_started) {
893 		printk(KERN_ERR "mesh: reselected with DMA started !\n");
894 		halt_dma(ms);
895 	}
896 	ms->current_req = NULL;
897 	ms->phase = dataing;
898 	ms->msgphase = msg_in;
899 	ms->n_msgout = 0;
900 	ms->last_n_msgout = 0;
901 	prev = ms->conn_tgt;
902 
903 	/*
904 	 * We seem to get abortive reselections sometimes.
905 	 */
906 	while ((in_8(&mr->bus_status1) & BS1_BSY) == 0) {
907 		static int mesh_aborted_resels;
908 		mesh_aborted_resels++;
909 		out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
910 		mesh_flush_io(mr);
911 		udelay(1);
912 		out_8(&mr->sequence, SEQ_ENBRESEL);
913 		mesh_flush_io(mr);
914 		udelay(5);
915 		dlog(ms, "extra resel err/exc/fc = %.6x",
916 		     MKWORD(0, mr->error, mr->exception, mr->fifo_count));
917 	}
918 	out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
919        	mesh_flush_io(mr);
920 	udelay(1);
921 	out_8(&mr->sequence, SEQ_ENBRESEL);
922        	mesh_flush_io(mr);
923 	udelay(1);
924 	out_8(&mr->sync_params, ASYNC_PARAMS);
925 
926 	/*
927 	 * Find out who reselected us.
928 	 */
929 	if (in_8(&mr->fifo_count) == 0) {
930 		printk(KERN_ERR "mesh: reselection but nothing in fifo?\n");
931 		ms->conn_tgt = ms->host->this_id;
932 		goto bogus;
933 	}
934 	/* get the last byte in the fifo */
935 	do {
936 		b = in_8(&mr->fifo);
937 		dlog(ms, "reseldata %x", b);
938 	} while (in_8(&mr->fifo_count));
939 	for (t = 0; t < 8; ++t)
940 		if ((b & (1 << t)) != 0 && t != ms->host->this_id)
941 			break;
942 	if (b != (1 << t) + (1 << ms->host->this_id)) {
943 		printk(KERN_ERR "mesh: bad reselection data %x\n", b);
944 		ms->conn_tgt = ms->host->this_id;
945 		goto bogus;
946 	}
947 
948 
949 	/*
950 	 * Set up to continue with that target's transfer.
951 	 */
952 	ms->conn_tgt = t;
953 	tp = &ms->tgts[t];
954 	out_8(&mr->sync_params, tp->sync_params);
955 	if (ALLOW_DEBUG(t)) {
956 		printk(KERN_DEBUG "mesh: reselected by target %d\n", t);
957 		printk(KERN_DEBUG "mesh: saved_ptr=%x goes_out=%d cmd=%p\n",
958 		       tp->saved_ptr, tp->data_goes_out, tp->current_req);
959 	}
960 	ms->current_req = tp->current_req;
961 	if (tp->current_req == NULL) {
962 		printk(KERN_ERR "mesh: reselected by tgt %d but no cmd!\n", t);
963 		goto bogus;
964 	}
965 	ms->data_ptr = tp->saved_ptr;
966 	dlog(ms, "resel prev tgt=%d", prev);
967 	dlog(ms, "resel err/exc=%.4x", MKWORD(0, 0, mr->error, mr->exception));
968 	start_phase(ms);
969 	return;
970 
971 bogus:
972 	dumplog(ms, ms->conn_tgt);
973 	dumpslog(ms);
974 	ms->data_ptr = 0;
975 	ms->aborting = 1;
976 	start_phase(ms);
977 }
978 
979 static void do_abort(struct mesh_state *ms)
980 {
981 	ms->msgout[0] = ABORT;
982 	ms->n_msgout = 1;
983 	ms->aborting = 1;
984 	ms->stat = DID_ABORT;
985 	dlog(ms, "abort", 0);
986 }
987 
988 static void handle_reset(struct mesh_state *ms)
989 {
990 	int tgt;
991 	struct mesh_target *tp;
992 	struct scsi_cmnd *cmd;
993 	volatile struct mesh_regs __iomem *mr = ms->mesh;
994 
995 	for (tgt = 0; tgt < 8; ++tgt) {
996 		tp = &ms->tgts[tgt];
997 		if ((cmd = tp->current_req) != NULL) {
998 			cmd->result = DID_RESET << 16;
999 			tp->current_req = NULL;
1000 			mesh_completed(ms, cmd);
1001 		}
1002 		ms->tgts[tgt].sdtr_state = do_sdtr;
1003 		ms->tgts[tgt].sync_params = ASYNC_PARAMS;
1004 	}
1005 	ms->current_req = NULL;
1006 	while ((cmd = ms->request_q) != NULL) {
1007 		ms->request_q = (struct scsi_cmnd *) cmd->host_scribble;
1008 		cmd->result = DID_RESET << 16;
1009 		mesh_completed(ms, cmd);
1010 	}
1011 	ms->phase = idle;
1012 	ms->msgphase = msg_none;
1013 	out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1014 	out_8(&mr->sequence, SEQ_FLUSHFIFO);
1015        	mesh_flush_io(mr);
1016 	udelay(1);
1017 	out_8(&mr->sync_params, ASYNC_PARAMS);
1018 	out_8(&mr->sequence, SEQ_ENBRESEL);
1019 }
1020 
1021 static irqreturn_t do_mesh_interrupt(int irq, void *dev_id)
1022 {
1023 	unsigned long flags;
1024 	struct mesh_state *ms = dev_id;
1025 	struct Scsi_Host *dev = ms->host;
1026 
1027 	spin_lock_irqsave(dev->host_lock, flags);
1028 	mesh_interrupt(ms);
1029 	spin_unlock_irqrestore(dev->host_lock, flags);
1030 	return IRQ_HANDLED;
1031 }
1032 
1033 static void handle_error(struct mesh_state *ms)
1034 {
1035 	int err, exc, count;
1036 	volatile struct mesh_regs __iomem *mr = ms->mesh;
1037 
1038 	err = in_8(&mr->error);
1039 	exc = in_8(&mr->exception);
1040 	out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1041 	dlog(ms, "error err/exc/fc/cl=%.8x",
1042 	     MKWORD(err, exc, mr->fifo_count, mr->count_lo));
1043 	if (err & ERR_SCSIRESET) {
1044 		/* SCSI bus was reset */
1045 		printk(KERN_INFO "mesh: SCSI bus reset detected: "
1046 		       "waiting for end...");
1047 		while ((in_8(&mr->bus_status1) & BS1_RST) != 0)
1048 			udelay(1);
1049 		printk("done\n");
1050 		handle_reset(ms);
1051 		/* request_q is empty, no point in mesh_start() */
1052 		return;
1053 	}
1054 	if (err & ERR_UNEXPDISC) {
1055 		/* Unexpected disconnect */
1056 		if (exc & EXC_RESELECTED) {
1057 			reselected(ms);
1058 			return;
1059 		}
1060 		if (!ms->aborting) {
1061 			printk(KERN_WARNING "mesh: target %d aborted\n",
1062 			       ms->conn_tgt);
1063 			dumplog(ms, ms->conn_tgt);
1064 			dumpslog(ms);
1065 		}
1066 		out_8(&mr->interrupt, INT_CMDDONE);
1067 		ms->stat = DID_ABORT;
1068 		mesh_done(ms, 1);
1069 		return;
1070 	}
1071 	if (err & ERR_PARITY) {
1072 		if (ms->msgphase == msg_in) {
1073 			printk(KERN_ERR "mesh: msg parity error, target %d\n",
1074 			       ms->conn_tgt);
1075 			ms->msgout[0] = MSG_PARITY_ERROR;
1076 			ms->n_msgout = 1;
1077 			ms->msgphase = msg_in_bad;
1078 			cmd_complete(ms);
1079 			return;
1080 		}
1081 		if (ms->stat == DID_OK) {
1082 			printk(KERN_ERR "mesh: parity error, target %d\n",
1083 			       ms->conn_tgt);
1084 			ms->stat = DID_PARITY;
1085 		}
1086 		count = (mr->count_hi << 8) + mr->count_lo;
1087 		if (count == 0) {
1088 			cmd_complete(ms);
1089 		} else {
1090 			/* reissue the data transfer command */
1091 			out_8(&mr->sequence, mr->sequence);
1092 		}
1093 		return;
1094 	}
1095 	if (err & ERR_SEQERR) {
1096 		if (exc & EXC_RESELECTED) {
1097 			/* This can happen if we issue a command to
1098 			   get the bus just after the target reselects us. */
1099 			static int mesh_resel_seqerr;
1100 			mesh_resel_seqerr++;
1101 			reselected(ms);
1102 			return;
1103 		}
1104 		if (exc == EXC_PHASEMM) {
1105 			static int mesh_phasemm_seqerr;
1106 			mesh_phasemm_seqerr++;
1107 			phase_mismatch(ms);
1108 			return;
1109 		}
1110 		printk(KERN_ERR "mesh: sequence error (err=%x exc=%x)\n",
1111 		       err, exc);
1112 	} else {
1113 		printk(KERN_ERR "mesh: unknown error %x (exc=%x)\n", err, exc);
1114 	}
1115 	mesh_dump_regs(ms);
1116 	dumplog(ms, ms->conn_tgt);
1117 	if (ms->phase > selecting && (in_8(&mr->bus_status1) & BS1_BSY)) {
1118 		/* try to do what the target wants */
1119 		do_abort(ms);
1120 		phase_mismatch(ms);
1121 		return;
1122 	}
1123 	ms->stat = DID_ERROR;
1124 	mesh_done(ms, 1);
1125 }
1126 
1127 static void handle_exception(struct mesh_state *ms)
1128 {
1129 	int exc;
1130 	volatile struct mesh_regs __iomem *mr = ms->mesh;
1131 
1132 	exc = in_8(&mr->exception);
1133 	out_8(&mr->interrupt, INT_EXCEPTION | INT_CMDDONE);
1134 	if (exc & EXC_RESELECTED) {
1135 		static int mesh_resel_exc;
1136 		mesh_resel_exc++;
1137 		reselected(ms);
1138 	} else if (exc == EXC_ARBLOST) {
1139 		printk(KERN_DEBUG "mesh: lost arbitration\n");
1140 		ms->stat = DID_BUS_BUSY;
1141 		mesh_done(ms, 1);
1142 	} else if (exc == EXC_SELTO) {
1143 		/* selection timed out */
1144 		ms->stat = DID_BAD_TARGET;
1145 		mesh_done(ms, 1);
1146 	} else if (exc == EXC_PHASEMM) {
1147 		/* target wants to do something different:
1148 		   find out what it wants and do it. */
1149 		phase_mismatch(ms);
1150 	} else {
1151 		printk(KERN_ERR "mesh: can't cope with exception %x\n", exc);
1152 		mesh_dump_regs(ms);
1153 		dumplog(ms, ms->conn_tgt);
1154 		do_abort(ms);
1155 		phase_mismatch(ms);
1156 	}
1157 }
1158 
1159 static void handle_msgin(struct mesh_state *ms)
1160 {
1161 	int i, code;
1162 	struct scsi_cmnd *cmd = ms->current_req;
1163 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
1164 
1165 	if (ms->n_msgin == 0)
1166 		return;
1167 	code = ms->msgin[0];
1168 	if (ALLOW_DEBUG(ms->conn_tgt)) {
1169 		printk(KERN_DEBUG "got %d message bytes:", ms->n_msgin);
1170 		for (i = 0; i < ms->n_msgin; ++i)
1171 			printk(" %x", ms->msgin[i]);
1172 		printk("\n");
1173 	}
1174 	dlog(ms, "msgin msg=%.8x",
1175 	     MKWORD(ms->n_msgin, code, ms->msgin[1], ms->msgin[2]));
1176 
1177 	ms->expect_reply = 0;
1178 	ms->n_msgout = 0;
1179 	if (ms->n_msgin < msgin_length(ms))
1180 		goto reject;
1181 	if (cmd)
1182 		cmd->SCp.Message = code;
1183 	switch (code) {
1184 	case COMMAND_COMPLETE:
1185 		break;
1186 	case EXTENDED_MESSAGE:
1187 		switch (ms->msgin[2]) {
1188 		case EXTENDED_MODIFY_DATA_POINTER:
1189 			ms->data_ptr += (ms->msgin[3] << 24) + ms->msgin[6]
1190 				+ (ms->msgin[4] << 16) + (ms->msgin[5] << 8);
1191 			break;
1192 		case EXTENDED_SDTR:
1193 			if (tp->sdtr_state != sdtr_sent) {
1194 				/* reply with an SDTR */
1195 				add_sdtr_msg(ms);
1196 				/* limit period to at least his value,
1197 				   offset to no more than his */
1198 				if (ms->msgout[3] < ms->msgin[3])
1199 					ms->msgout[3] = ms->msgin[3];
1200 				if (ms->msgout[4] > ms->msgin[4])
1201 					ms->msgout[4] = ms->msgin[4];
1202 				set_sdtr(ms, ms->msgout[3], ms->msgout[4]);
1203 				ms->msgphase = msg_out;
1204 			} else {
1205 				set_sdtr(ms, ms->msgin[3], ms->msgin[4]);
1206 			}
1207 			break;
1208 		default:
1209 			goto reject;
1210 		}
1211 		break;
1212 	case SAVE_POINTERS:
1213 		tp->saved_ptr = ms->data_ptr;
1214 		break;
1215 	case RESTORE_POINTERS:
1216 		ms->data_ptr = tp->saved_ptr;
1217 		break;
1218 	case DISCONNECT:
1219 		ms->phase = disconnecting;
1220 		break;
1221 	case ABORT:
1222 		break;
1223 	case MESSAGE_REJECT:
1224 		if (tp->sdtr_state == sdtr_sent)
1225 			set_sdtr(ms, 0, 0);
1226 		break;
1227 	case NOP:
1228 		break;
1229 	default:
1230 		if (IDENTIFY_BASE <= code && code <= IDENTIFY_BASE + 7) {
1231 			if (cmd == NULL) {
1232 				do_abort(ms);
1233 				ms->msgphase = msg_out;
1234 			} else if (code != cmd->device->lun + IDENTIFY_BASE) {
1235 				printk(KERN_WARNING "mesh: lun mismatch "
1236 				       "(%d != %d) on reselection from "
1237 				       "target %d\n", code - IDENTIFY_BASE,
1238 				       cmd->device->lun, ms->conn_tgt);
1239 			}
1240 			break;
1241 		}
1242 		goto reject;
1243 	}
1244 	return;
1245 
1246  reject:
1247 	printk(KERN_WARNING "mesh: rejecting message from target %d:",
1248 	       ms->conn_tgt);
1249 	for (i = 0; i < ms->n_msgin; ++i)
1250 		printk(" %x", ms->msgin[i]);
1251 	printk("\n");
1252 	ms->msgout[0] = MESSAGE_REJECT;
1253 	ms->n_msgout = 1;
1254 	ms->msgphase = msg_out;
1255 }
1256 
1257 /*
1258  * Set up DMA commands for transferring data.
1259  */
1260 static void set_dma_cmds(struct mesh_state *ms, struct scsi_cmnd *cmd)
1261 {
1262 	int i, dma_cmd, total, off, dtot;
1263 	struct scatterlist *scl;
1264 	struct dbdma_cmd *dcmds;
1265 
1266 	dma_cmd = ms->tgts[ms->conn_tgt].data_goes_out?
1267 		OUTPUT_MORE: INPUT_MORE;
1268 	dcmds = ms->dma_cmds;
1269 	dtot = 0;
1270 	if (cmd) {
1271 		int nseg;
1272 
1273 		cmd->SCp.this_residual = scsi_bufflen(cmd);
1274 
1275 		nseg = scsi_dma_map(cmd);
1276 		BUG_ON(nseg < 0);
1277 
1278 		if (nseg) {
1279 			total = 0;
1280 			off = ms->data_ptr;
1281 
1282 			scsi_for_each_sg(cmd, scl, nseg, i) {
1283 				u32 dma_addr = sg_dma_address(scl);
1284 				u32 dma_len = sg_dma_len(scl);
1285 
1286 				total += scl->length;
1287 				if (off >= dma_len) {
1288 					off -= dma_len;
1289 					continue;
1290 				}
1291 				if (dma_len > 0xffff)
1292 					panic("mesh: scatterlist element >= 64k");
1293 				st_le16(&dcmds->req_count, dma_len - off);
1294 				st_le16(&dcmds->command, dma_cmd);
1295 				st_le32(&dcmds->phy_addr, dma_addr + off);
1296 				dcmds->xfer_status = 0;
1297 				++dcmds;
1298 				dtot += dma_len - off;
1299 				off = 0;
1300 			}
1301 		}
1302 	}
1303 	if (dtot == 0) {
1304 		/* Either the target has overrun our buffer,
1305 		   or the caller didn't provide a buffer. */
1306 		static char mesh_extra_buf[64];
1307 
1308 		dtot = sizeof(mesh_extra_buf);
1309 		st_le16(&dcmds->req_count, dtot);
1310 		st_le32(&dcmds->phy_addr, virt_to_phys(mesh_extra_buf));
1311 		dcmds->xfer_status = 0;
1312 		++dcmds;
1313 	}
1314 	dma_cmd += OUTPUT_LAST - OUTPUT_MORE;
1315 	st_le16(&dcmds[-1].command, dma_cmd);
1316 	memset(dcmds, 0, sizeof(*dcmds));
1317 	st_le16(&dcmds->command, DBDMA_STOP);
1318 	ms->dma_count = dtot;
1319 }
1320 
1321 static void halt_dma(struct mesh_state *ms)
1322 {
1323 	volatile struct dbdma_regs __iomem *md = ms->dma;
1324 	volatile struct mesh_regs __iomem *mr = ms->mesh;
1325 	struct scsi_cmnd *cmd = ms->current_req;
1326 	int t, nb;
1327 
1328 	if (!ms->tgts[ms->conn_tgt].data_goes_out) {
1329 		/* wait a little while until the fifo drains */
1330 		t = 50;
1331 		while (t > 0 && in_8(&mr->fifo_count) != 0
1332 		       && (in_le32(&md->status) & ACTIVE) != 0) {
1333 			--t;
1334 			udelay(1);
1335 		}
1336 	}
1337 	out_le32(&md->control, RUN << 16);	/* turn off RUN bit */
1338 	nb = (mr->count_hi << 8) + mr->count_lo;
1339 	dlog(ms, "halt_dma fc/count=%.6x",
1340 	     MKWORD(0, mr->fifo_count, 0, nb));
1341 	if (ms->tgts[ms->conn_tgt].data_goes_out)
1342 		nb += mr->fifo_count;
1343 	/* nb is the number of bytes not yet transferred
1344 	   to/from the target. */
1345 	ms->data_ptr -= nb;
1346 	dlog(ms, "data_ptr %x", ms->data_ptr);
1347 	if (ms->data_ptr < 0) {
1348 		printk(KERN_ERR "mesh: halt_dma: data_ptr=%d (nb=%d, ms=%p)\n",
1349 		       ms->data_ptr, nb, ms);
1350 		ms->data_ptr = 0;
1351 #ifdef MESH_DBG
1352 		dumplog(ms, ms->conn_tgt);
1353 		dumpslog(ms);
1354 #endif /* MESH_DBG */
1355 	} else if (cmd && scsi_bufflen(cmd) &&
1356 		   ms->data_ptr > scsi_bufflen(cmd)) {
1357 		printk(KERN_DEBUG "mesh: target %d overrun, "
1358 		       "data_ptr=%x total=%x goes_out=%d\n",
1359 		       ms->conn_tgt, ms->data_ptr, scsi_bufflen(cmd),
1360 		       ms->tgts[ms->conn_tgt].data_goes_out);
1361 	}
1362 	scsi_dma_unmap(cmd);
1363 	ms->dma_started = 0;
1364 }
1365 
1366 static void phase_mismatch(struct mesh_state *ms)
1367 {
1368 	volatile struct mesh_regs __iomem *mr = ms->mesh;
1369 	int phase;
1370 
1371 	dlog(ms, "phasemm ch/cl/seq/fc=%.8x",
1372 	     MKWORD(mr->count_hi, mr->count_lo, mr->sequence, mr->fifo_count));
1373 	phase = in_8(&mr->bus_status0) & BS0_PHASE;
1374 	if (ms->msgphase == msg_out_xxx && phase == BP_MSGOUT) {
1375 		/* output the last byte of the message, without ATN */
1376 		out_8(&mr->count_lo, 1);
1377 		out_8(&mr->sequence, SEQ_MSGOUT + use_active_neg);
1378 		mesh_flush_io(mr);
1379 		udelay(1);
1380 		out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]);
1381 		ms->msgphase = msg_out_last;
1382 		return;
1383 	}
1384 
1385 	if (ms->msgphase == msg_in) {
1386 		get_msgin(ms);
1387 		if (ms->n_msgin)
1388 			handle_msgin(ms);
1389 	}
1390 
1391 	if (ms->dma_started)
1392 		halt_dma(ms);
1393 	if (mr->fifo_count) {
1394 		out_8(&mr->sequence, SEQ_FLUSHFIFO);
1395 		mesh_flush_io(mr);
1396 		udelay(1);
1397 	}
1398 
1399 	ms->msgphase = msg_none;
1400 	switch (phase) {
1401 	case BP_DATAIN:
1402 		ms->tgts[ms->conn_tgt].data_goes_out = 0;
1403 		ms->phase = dataing;
1404 		break;
1405 	case BP_DATAOUT:
1406 		ms->tgts[ms->conn_tgt].data_goes_out = 1;
1407 		ms->phase = dataing;
1408 		break;
1409 	case BP_COMMAND:
1410 		ms->phase = commanding;
1411 		break;
1412 	case BP_STATUS:
1413 		ms->phase = statusing;
1414 		break;
1415 	case BP_MSGIN:
1416 		ms->msgphase = msg_in;
1417 		ms->n_msgin = 0;
1418 		break;
1419 	case BP_MSGOUT:
1420 		ms->msgphase = msg_out;
1421 		if (ms->n_msgout == 0) {
1422 			if (ms->aborting) {
1423 				do_abort(ms);
1424 			} else {
1425 				if (ms->last_n_msgout == 0) {
1426 					printk(KERN_DEBUG
1427 					       "mesh: no msg to repeat\n");
1428 					ms->msgout[0] = NOP;
1429 					ms->last_n_msgout = 1;
1430 				}
1431 				ms->n_msgout = ms->last_n_msgout;
1432 			}
1433 		}
1434 		break;
1435 	default:
1436 		printk(KERN_DEBUG "mesh: unknown scsi phase %x\n", phase);
1437 		ms->stat = DID_ERROR;
1438 		mesh_done(ms, 1);
1439 		return;
1440 	}
1441 
1442 	start_phase(ms);
1443 }
1444 
1445 static void cmd_complete(struct mesh_state *ms)
1446 {
1447 	volatile struct mesh_regs __iomem *mr = ms->mesh;
1448 	struct scsi_cmnd *cmd = ms->current_req;
1449 	struct mesh_target *tp = &ms->tgts[ms->conn_tgt];
1450 	int seq, n, t;
1451 
1452 	dlog(ms, "cmd_complete fc=%x", mr->fifo_count);
1453 	seq = use_active_neg + (ms->n_msgout? SEQ_ATN: 0);
1454 	switch (ms->msgphase) {
1455 	case msg_out_xxx:
1456 		/* huh?  we expected a phase mismatch */
1457 		ms->n_msgin = 0;
1458 		ms->msgphase = msg_in;
1459 		/* fall through */
1460 
1461 	case msg_in:
1462 		/* should have some message bytes in fifo */
1463 		get_msgin(ms);
1464 		n = msgin_length(ms);
1465 		if (ms->n_msgin < n) {
1466 			out_8(&mr->count_lo, n - ms->n_msgin);
1467 			out_8(&mr->sequence, SEQ_MSGIN + seq);
1468 		} else {
1469 			ms->msgphase = msg_none;
1470 			handle_msgin(ms);
1471 			start_phase(ms);
1472 		}
1473 		break;
1474 
1475 	case msg_in_bad:
1476 		out_8(&mr->sequence, SEQ_FLUSHFIFO);
1477 		mesh_flush_io(mr);
1478 		udelay(1);
1479 		out_8(&mr->count_lo, 1);
1480 		out_8(&mr->sequence, SEQ_MSGIN + SEQ_ATN + use_active_neg);
1481 		break;
1482 
1483 	case msg_out:
1484 		/*
1485 		 * To get the right timing on ATN wrt ACK, we have
1486 		 * to get the MESH to drop ACK, wait until REQ gets
1487 		 * asserted, then drop ATN.  To do this we first
1488 		 * issue a SEQ_MSGOUT with ATN and wait for REQ,
1489 		 * then change the command to a SEQ_MSGOUT w/o ATN.
1490 		 * If we don't see REQ in a reasonable time, we
1491 		 * change the command to SEQ_MSGIN with ATN,
1492 		 * wait for the phase mismatch interrupt, then
1493 		 * issue the SEQ_MSGOUT without ATN.
1494 		 */
1495 		out_8(&mr->count_lo, 1);
1496 		out_8(&mr->sequence, SEQ_MSGOUT + use_active_neg + SEQ_ATN);
1497 		t = 30;		/* wait up to 30us */
1498 		while ((in_8(&mr->bus_status0) & BS0_REQ) == 0 && --t >= 0)
1499 			udelay(1);
1500 		dlog(ms, "last_mbyte err/exc/fc/cl=%.8x",
1501 		     MKWORD(mr->error, mr->exception,
1502 			    mr->fifo_count, mr->count_lo));
1503 		if (in_8(&mr->interrupt) & (INT_ERROR | INT_EXCEPTION)) {
1504 			/* whoops, target didn't do what we expected */
1505 			ms->last_n_msgout = ms->n_msgout;
1506 			ms->n_msgout = 0;
1507 			if (in_8(&mr->interrupt) & INT_ERROR) {
1508 				printk(KERN_ERR "mesh: error %x in msg_out\n",
1509 				       in_8(&mr->error));
1510 				handle_error(ms);
1511 				return;
1512 			}
1513 			if (in_8(&mr->exception) != EXC_PHASEMM)
1514 				printk(KERN_ERR "mesh: exc %x in msg_out\n",
1515 				       in_8(&mr->exception));
1516 			else
1517 				printk(KERN_DEBUG "mesh: bs0=%x in msg_out\n",
1518 				       in_8(&mr->bus_status0));
1519 			handle_exception(ms);
1520 			return;
1521 		}
1522 		if (in_8(&mr->bus_status0) & BS0_REQ) {
1523 			out_8(&mr->sequence, SEQ_MSGOUT + use_active_neg);
1524 			mesh_flush_io(mr);
1525 			udelay(1);
1526 			out_8(&mr->fifo, ms->msgout[ms->n_msgout-1]);
1527 			ms->msgphase = msg_out_last;
1528 		} else {
1529 			out_8(&mr->sequence, SEQ_MSGIN + use_active_neg + SEQ_ATN);
1530 			ms->msgphase = msg_out_xxx;
1531 		}
1532 		break;
1533 
1534 	case msg_out_last:
1535 		ms->last_n_msgout = ms->n_msgout;
1536 		ms->n_msgout = 0;
1537 		ms->msgphase = ms->expect_reply? msg_in: msg_none;
1538 		start_phase(ms);
1539 		break;
1540 
1541 	case msg_none:
1542 		switch (ms->phase) {
1543 		case idle:
1544 			printk(KERN_ERR "mesh: interrupt in idle phase?\n");
1545 			dumpslog(ms);
1546 			return;
1547 		case selecting:
1548 			dlog(ms, "Selecting phase at command completion",0);
1549 			ms->msgout[0] = IDENTIFY(ALLOW_RESEL(ms->conn_tgt),
1550 						 (cmd? cmd->device->lun: 0));
1551 			ms->n_msgout = 1;
1552 			ms->expect_reply = 0;
1553 			if (ms->aborting) {
1554 				ms->msgout[0] = ABORT;
1555 				ms->n_msgout++;
1556 			} else if (tp->sdtr_state == do_sdtr) {
1557 				/* add SDTR message */
1558 				add_sdtr_msg(ms);
1559 				ms->expect_reply = 1;
1560 				tp->sdtr_state = sdtr_sent;
1561 			}
1562 			ms->msgphase = msg_out;
1563 			/*
1564 			 * We need to wait for REQ before dropping ATN.
1565 			 * We wait for at most 30us, then fall back to
1566 			 * a scheme where we issue a SEQ_COMMAND with ATN,
1567 			 * which will give us a phase mismatch interrupt
1568 			 * when REQ does come, and then we send the message.
1569 			 */
1570 			t = 230;		/* wait up to 230us */
1571 			while ((in_8(&mr->bus_status0) & BS0_REQ) == 0) {
1572 				if (--t < 0) {
1573 					dlog(ms, "impatient for req", ms->n_msgout);
1574 					ms->msgphase = msg_none;
1575 					break;
1576 				}
1577 				udelay(1);
1578 			}
1579 			break;
1580 		case dataing:
1581 			if (ms->dma_count != 0) {
1582 				start_phase(ms);
1583 				return;
1584 			}
1585 			/*
1586 			 * We can get a phase mismatch here if the target
1587 			 * changes to the status phase, even though we have
1588 			 * had a command complete interrupt.  Then, if we
1589 			 * issue the SEQ_STATUS command, we'll get a sequence
1590 			 * error interrupt.  Which isn't so bad except that
1591 			 * occasionally the mesh actually executes the
1592 			 * SEQ_STATUS *as well as* giving us the sequence
1593 			 * error and phase mismatch exception.
1594 			 */
1595 			out_8(&mr->sequence, 0);
1596 			out_8(&mr->interrupt,
1597 			      INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1598 			halt_dma(ms);
1599 			break;
1600 		case statusing:
1601 			if (cmd) {
1602 				cmd->SCp.Status = mr->fifo;
1603 				if (DEBUG_TARGET(cmd))
1604 					printk(KERN_DEBUG "mesh: status is %x\n",
1605 					       cmd->SCp.Status);
1606 			}
1607 			ms->msgphase = msg_in;
1608 			break;
1609 		case busfreeing:
1610 			mesh_done(ms, 1);
1611 			return;
1612 		case disconnecting:
1613 			ms->current_req = NULL;
1614 			ms->phase = idle;
1615 			mesh_start(ms);
1616 			return;
1617 		default:
1618 			break;
1619 		}
1620 		++ms->phase;
1621 		start_phase(ms);
1622 		break;
1623 	}
1624 }
1625 
1626 
1627 /*
1628  * Called by midlayer with host locked to queue a new
1629  * request
1630  */
1631 static int mesh_queue(struct scsi_cmnd *cmd, void (*done)(struct scsi_cmnd *))
1632 {
1633 	struct mesh_state *ms;
1634 
1635 	cmd->scsi_done = done;
1636 	cmd->host_scribble = NULL;
1637 
1638 	ms = (struct mesh_state *) cmd->device->host->hostdata;
1639 
1640 	if (ms->request_q == NULL)
1641 		ms->request_q = cmd;
1642 	else
1643 		ms->request_qtail->host_scribble = (void *) cmd;
1644 	ms->request_qtail = cmd;
1645 
1646 	if (ms->phase == idle)
1647 		mesh_start(ms);
1648 
1649 	return 0;
1650 }
1651 
1652 /*
1653  * Called to handle interrupts, either call by the interrupt
1654  * handler (do_mesh_interrupt) or by other functions in
1655  * exceptional circumstances
1656  */
1657 static void mesh_interrupt(struct mesh_state *ms)
1658 {
1659 	volatile struct mesh_regs __iomem *mr = ms->mesh;
1660 	int intr;
1661 
1662 #if 0
1663 	if (ALLOW_DEBUG(ms->conn_tgt))
1664 		printk(KERN_DEBUG "mesh_intr, bs0=%x int=%x exc=%x err=%x "
1665 		       "phase=%d msgphase=%d\n", mr->bus_status0,
1666 		       mr->interrupt, mr->exception, mr->error,
1667 		       ms->phase, ms->msgphase);
1668 #endif
1669 	while ((intr = in_8(&mr->interrupt)) != 0) {
1670 		dlog(ms, "interrupt intr/err/exc/seq=%.8x",
1671 		     MKWORD(intr, mr->error, mr->exception, mr->sequence));
1672 		if (intr & INT_ERROR) {
1673 			handle_error(ms);
1674 		} else if (intr & INT_EXCEPTION) {
1675 			handle_exception(ms);
1676 		} else if (intr & INT_CMDDONE) {
1677 			out_8(&mr->interrupt, INT_CMDDONE);
1678 			cmd_complete(ms);
1679 		}
1680 	}
1681 }
1682 
1683 /* Todo: here we can at least try to remove the command from the
1684  * queue if it isn't connected yet, and for pending command, assert
1685  * ATN until the bus gets freed.
1686  */
1687 static int mesh_abort(struct scsi_cmnd *cmd)
1688 {
1689 	struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata;
1690 
1691 	printk(KERN_DEBUG "mesh_abort(%p)\n", cmd);
1692 	mesh_dump_regs(ms);
1693 	dumplog(ms, cmd->device->id);
1694 	dumpslog(ms);
1695 	return FAILED;
1696 }
1697 
1698 /*
1699  * Called by the midlayer with the lock held to reset the
1700  * SCSI host and bus.
1701  * The midlayer will wait for devices to come back, we don't need
1702  * to do that ourselves
1703  */
1704 static int mesh_host_reset(struct scsi_cmnd *cmd)
1705 {
1706 	struct mesh_state *ms = (struct mesh_state *) cmd->device->host->hostdata;
1707 	volatile struct mesh_regs __iomem *mr = ms->mesh;
1708 	volatile struct dbdma_regs __iomem *md = ms->dma;
1709 	unsigned long flags;
1710 
1711 	printk(KERN_DEBUG "mesh_host_reset\n");
1712 
1713 	spin_lock_irqsave(ms->host->host_lock, flags);
1714 
1715 	/* Reset the controller & dbdma channel */
1716 	out_le32(&md->control, (RUN|PAUSE|FLUSH|WAKE) << 16);	/* stop dma */
1717 	out_8(&mr->exception, 0xff);	/* clear all exception bits */
1718 	out_8(&mr->error, 0xff);	/* clear all error bits */
1719 	out_8(&mr->sequence, SEQ_RESETMESH);
1720        	mesh_flush_io(mr);
1721 	udelay(1);
1722 	out_8(&mr->intr_mask, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1723 	out_8(&mr->source_id, ms->host->this_id);
1724 	out_8(&mr->sel_timeout, 25);	/* 250ms */
1725 	out_8(&mr->sync_params, ASYNC_PARAMS);
1726 
1727 	/* Reset the bus */
1728 	out_8(&mr->bus_status1, BS1_RST);	/* assert RST */
1729        	mesh_flush_io(mr);
1730 	udelay(30);			/* leave it on for >= 25us */
1731 	out_8(&mr->bus_status1, 0);	/* negate RST */
1732 
1733 	/* Complete pending commands */
1734 	handle_reset(ms);
1735 
1736 	spin_unlock_irqrestore(ms->host->host_lock, flags);
1737 	return SUCCESS;
1738 }
1739 
1740 static void set_mesh_power(struct mesh_state *ms, int state)
1741 {
1742 	if (!machine_is(powermac))
1743 		return;
1744 	if (state) {
1745 		pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 1);
1746 		msleep(200);
1747 	} else {
1748 		pmac_call_feature(PMAC_FTR_MESH_ENABLE, macio_get_of_node(ms->mdev), 0, 0);
1749 		msleep(10);
1750 	}
1751 }
1752 
1753 
1754 #ifdef CONFIG_PM
1755 static int mesh_suspend(struct macio_dev *mdev, pm_message_t mesg)
1756 {
1757 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
1758 	unsigned long flags;
1759 
1760 	switch (mesg.event) {
1761 	case PM_EVENT_SUSPEND:
1762 	case PM_EVENT_FREEZE:
1763 		break;
1764 	default:
1765 		return 0;
1766 	}
1767 	if (mesg.event == mdev->ofdev.dev.power.power_state.event)
1768 		return 0;
1769 
1770 	scsi_block_requests(ms->host);
1771 	spin_lock_irqsave(ms->host->host_lock, flags);
1772 	while(ms->phase != idle) {
1773 		spin_unlock_irqrestore(ms->host->host_lock, flags);
1774 		msleep(10);
1775 		spin_lock_irqsave(ms->host->host_lock, flags);
1776 	}
1777 	ms->phase = sleeping;
1778 	spin_unlock_irqrestore(ms->host->host_lock, flags);
1779 	disable_irq(ms->meshintr);
1780 	set_mesh_power(ms, 0);
1781 
1782 	mdev->ofdev.dev.power.power_state = mesg;
1783 
1784 	return 0;
1785 }
1786 
1787 static int mesh_resume(struct macio_dev *mdev)
1788 {
1789 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
1790 	unsigned long flags;
1791 
1792 	if (mdev->ofdev.dev.power.power_state.event == PM_EVENT_ON)
1793 		return 0;
1794 
1795 	set_mesh_power(ms, 1);
1796 	mesh_init(ms);
1797 	spin_lock_irqsave(ms->host->host_lock, flags);
1798 	mesh_start(ms);
1799 	spin_unlock_irqrestore(ms->host->host_lock, flags);
1800 	enable_irq(ms->meshintr);
1801 	scsi_unblock_requests(ms->host);
1802 
1803 	mdev->ofdev.dev.power.power_state.event = PM_EVENT_ON;
1804 
1805 	return 0;
1806 }
1807 
1808 #endif /* CONFIG_PM */
1809 
1810 /*
1811  * If we leave drives set for synchronous transfers (especially
1812  * CDROMs), and reboot to MacOS, it gets confused, poor thing.
1813  * So, on reboot we reset the SCSI bus.
1814  */
1815 static int mesh_shutdown(struct macio_dev *mdev)
1816 {
1817 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
1818 	volatile struct mesh_regs __iomem *mr;
1819 	unsigned long flags;
1820 
1821        	printk(KERN_INFO "resetting MESH scsi bus(es)\n");
1822 	spin_lock_irqsave(ms->host->host_lock, flags);
1823        	mr = ms->mesh;
1824 	out_8(&mr->intr_mask, 0);
1825 	out_8(&mr->interrupt, INT_ERROR | INT_EXCEPTION | INT_CMDDONE);
1826 	out_8(&mr->bus_status1, BS1_RST);
1827 	mesh_flush_io(mr);
1828 	udelay(30);
1829 	out_8(&mr->bus_status1, 0);
1830 	spin_unlock_irqrestore(ms->host->host_lock, flags);
1831 
1832 	return 0;
1833 }
1834 
1835 static struct scsi_host_template mesh_template = {
1836 	.proc_name			= "mesh",
1837 	.name				= "MESH",
1838 	.queuecommand			= mesh_queue,
1839 	.eh_abort_handler		= mesh_abort,
1840 	.eh_host_reset_handler		= mesh_host_reset,
1841 	.can_queue			= 20,
1842 	.this_id			= 7,
1843 	.sg_tablesize			= SG_ALL,
1844 	.cmd_per_lun			= 2,
1845 	.use_clustering			= DISABLE_CLUSTERING,
1846 };
1847 
1848 static int mesh_probe(struct macio_dev *mdev, const struct of_device_id *match)
1849 {
1850 	struct device_node *mesh = macio_get_of_node(mdev);
1851 	struct pci_dev* pdev = macio_get_pci_dev(mdev);
1852 	int tgt, minper;
1853 	const int *cfp;
1854 	struct mesh_state *ms;
1855 	struct Scsi_Host *mesh_host;
1856 	void *dma_cmd_space;
1857 	dma_addr_t dma_cmd_bus;
1858 
1859 	switch (mdev->bus->chip->type) {
1860 	case macio_heathrow:
1861 	case macio_gatwick:
1862 	case macio_paddington:
1863 		use_active_neg = 0;
1864 		break;
1865 	default:
1866 		use_active_neg = SEQ_ACTIVE_NEG;
1867 	}
1868 
1869 	if (macio_resource_count(mdev) != 2 || macio_irq_count(mdev) != 2) {
1870        		printk(KERN_ERR "mesh: expected 2 addrs and 2 intrs"
1871 	       	       " (got %d,%d)\n", macio_resource_count(mdev),
1872 		       macio_irq_count(mdev));
1873 		return -ENODEV;
1874 	}
1875 
1876 	if (macio_request_resources(mdev, "mesh") != 0) {
1877        		printk(KERN_ERR "mesh: unable to request memory resources");
1878 		return -EBUSY;
1879 	}
1880        	mesh_host = scsi_host_alloc(&mesh_template, sizeof(struct mesh_state));
1881 	if (mesh_host == NULL) {
1882 		printk(KERN_ERR "mesh: couldn't register host");
1883 		goto out_release;
1884 	}
1885 
1886 	/* Old junk for root discovery, that will die ultimately */
1887 #if !defined(MODULE)
1888        	note_scsi_host(mesh, mesh_host);
1889 #endif
1890 
1891 	mesh_host->base = macio_resource_start(mdev, 0);
1892 	mesh_host->irq = macio_irq(mdev, 0);
1893        	ms = (struct mesh_state *) mesh_host->hostdata;
1894 	macio_set_drvdata(mdev, ms);
1895 	ms->host = mesh_host;
1896 	ms->mdev = mdev;
1897 	ms->pdev = pdev;
1898 
1899 	ms->mesh = ioremap(macio_resource_start(mdev, 0), 0x1000);
1900 	if (ms->mesh == NULL) {
1901 		printk(KERN_ERR "mesh: can't map registers\n");
1902 		goto out_free;
1903 	}
1904 	ms->dma = ioremap(macio_resource_start(mdev, 1), 0x1000);
1905 	if (ms->dma == NULL) {
1906 		printk(KERN_ERR "mesh: can't map registers\n");
1907 		iounmap(ms->mesh);
1908 		goto out_free;
1909 	}
1910 
1911        	ms->meshintr = macio_irq(mdev, 0);
1912        	ms->dmaintr = macio_irq(mdev, 1);
1913 
1914        	/* Space for dma command list: +1 for stop command,
1915        	 * +1 to allow for aligning.
1916 	 */
1917 	ms->dma_cmd_size = (mesh_host->sg_tablesize + 2) * sizeof(struct dbdma_cmd);
1918 
1919 	/* We use the PCI APIs for now until the generic one gets fixed
1920 	 * enough or until we get some macio-specific versions
1921 	 */
1922 	dma_cmd_space = pci_alloc_consistent(macio_get_pci_dev(mdev),
1923 					     ms->dma_cmd_size,
1924 					     &dma_cmd_bus);
1925 	if (dma_cmd_space == NULL) {
1926 		printk(KERN_ERR "mesh: can't allocate DMA table\n");
1927 		goto out_unmap;
1928 	}
1929 	memset(dma_cmd_space, 0, ms->dma_cmd_size);
1930 
1931 	ms->dma_cmds = (struct dbdma_cmd *) DBDMA_ALIGN(dma_cmd_space);
1932        	ms->dma_cmd_space = dma_cmd_space;
1933 	ms->dma_cmd_bus = dma_cmd_bus + ((unsigned long)ms->dma_cmds)
1934 		- (unsigned long)dma_cmd_space;
1935 	ms->current_req = NULL;
1936        	for (tgt = 0; tgt < 8; ++tgt) {
1937 	       	ms->tgts[tgt].sdtr_state = do_sdtr;
1938 	       	ms->tgts[tgt].sync_params = ASYNC_PARAMS;
1939 	       	ms->tgts[tgt].current_req = NULL;
1940        	}
1941 
1942 	if ((cfp = of_get_property(mesh, "clock-frequency", NULL)))
1943        		ms->clk_freq = *cfp;
1944 	else {
1945        		printk(KERN_INFO "mesh: assuming 50MHz clock frequency\n");
1946 	       	ms->clk_freq = 50000000;
1947        	}
1948 
1949        	/* The maximum sync rate is clock / 5; increase
1950        	 * mesh_sync_period if necessary.
1951 	 */
1952 	minper = 1000000000 / (ms->clk_freq / 5); /* ns */
1953 	if (mesh_sync_period < minper)
1954 		mesh_sync_period = minper;
1955 
1956 	/* Power up the chip */
1957 	set_mesh_power(ms, 1);
1958 
1959 	/* Set it up */
1960        	mesh_init(ms);
1961 
1962 	/* Request interrupt */
1963        	if (request_irq(ms->meshintr, do_mesh_interrupt, 0, "MESH", ms)) {
1964 	       	printk(KERN_ERR "MESH: can't get irq %d\n", ms->meshintr);
1965 		goto out_shutdown;
1966 	}
1967 
1968 	/* Add scsi host & scan */
1969 	if (scsi_add_host(mesh_host, &mdev->ofdev.dev))
1970 		goto out_release_irq;
1971 	scsi_scan_host(mesh_host);
1972 
1973 	return 0;
1974 
1975  out_release_irq:
1976 	free_irq(ms->meshintr, ms);
1977  out_shutdown:
1978 	/* shutdown & reset bus in case of error or macos can be confused
1979 	 * at reboot if the bus was set to synchronous mode already
1980 	 */
1981 	mesh_shutdown(mdev);
1982 	set_mesh_power(ms, 0);
1983 	pci_free_consistent(macio_get_pci_dev(mdev), ms->dma_cmd_size,
1984 			    ms->dma_cmd_space, ms->dma_cmd_bus);
1985  out_unmap:
1986 	iounmap(ms->dma);
1987 	iounmap(ms->mesh);
1988  out_free:
1989 	scsi_host_put(mesh_host);
1990  out_release:
1991 	macio_release_resources(mdev);
1992 
1993 	return -ENODEV;
1994 }
1995 
1996 static int mesh_remove(struct macio_dev *mdev)
1997 {
1998 	struct mesh_state *ms = (struct mesh_state *)macio_get_drvdata(mdev);
1999 	struct Scsi_Host *mesh_host = ms->host;
2000 
2001 	scsi_remove_host(mesh_host);
2002 
2003 	free_irq(ms->meshintr, ms);
2004 
2005 	/* Reset scsi bus */
2006 	mesh_shutdown(mdev);
2007 
2008 	/* Shut down chip & termination */
2009 	set_mesh_power(ms, 0);
2010 
2011 	/* Unmap registers & dma controller */
2012 	iounmap(ms->mesh);
2013        	iounmap(ms->dma);
2014 
2015 	/* Free DMA commands memory */
2016 	pci_free_consistent(macio_get_pci_dev(mdev), ms->dma_cmd_size,
2017 			    ms->dma_cmd_space, ms->dma_cmd_bus);
2018 
2019 	/* Release memory resources */
2020 	macio_release_resources(mdev);
2021 
2022 	scsi_host_put(mesh_host);
2023 
2024 	return 0;
2025 }
2026 
2027 
2028 static struct of_device_id mesh_match[] =
2029 {
2030 	{
2031 	.name 		= "mesh",
2032 	},
2033 	{
2034 	.type		= "scsi",
2035 	.compatible	= "chrp,mesh0"
2036 	},
2037 	{},
2038 };
2039 MODULE_DEVICE_TABLE (of, mesh_match);
2040 
2041 static struct macio_driver mesh_driver =
2042 {
2043 	.name 		= "mesh",
2044 	.match_table	= mesh_match,
2045 	.probe		= mesh_probe,
2046 	.remove		= mesh_remove,
2047 	.shutdown	= mesh_shutdown,
2048 #ifdef CONFIG_PM
2049 	.suspend	= mesh_suspend,
2050 	.resume		= mesh_resume,
2051 #endif
2052 };
2053 
2054 
2055 static int __init init_mesh(void)
2056 {
2057 
2058 	/* Calculate sync rate from module parameters */
2059 	if (sync_rate > 10)
2060 		sync_rate = 10;
2061 	if (sync_rate > 0) {
2062 		printk(KERN_INFO "mesh: configured for synchronous %d MB/s\n", sync_rate);
2063 		mesh_sync_period = 1000 / sync_rate;	/* ns */
2064 		mesh_sync_offset = 15;
2065 	} else
2066 		printk(KERN_INFO "mesh: configured for asynchronous\n");
2067 
2068 	return macio_register_driver(&mesh_driver);
2069 }
2070 
2071 static void __exit exit_mesh(void)
2072 {
2073 	return macio_unregister_driver(&mesh_driver);
2074 }
2075 
2076 module_init(init_mesh);
2077 module_exit(exit_mesh);
2078