xref: /openbmc/linux/drivers/usb/host/isp1362-hcd.c (revision dd5e5554)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * ISP1362 HCD (Host Controller Driver) for USB.
4  *
5  * Copyright (C) 2005 Lothar Wassmann <LW@KARO-electronics.de>
6  *
7  * Derived from the SL811 HCD, rewritten for ISP116x.
8  * Copyright (C) 2005 Olav Kongas <ok@artecdesign.ee>
9  *
10  * Portions:
11  * Copyright (C) 2004 Psion Teklogix (for NetBook PRO)
12  * Copyright (C) 2004 David Brownell
13  */
14 
15 /*
16  * The ISP1362 chip requires a large delay (300ns and 462ns) between
17  * accesses to the address and data register.
18  * The following timing options exist:
19  *
20  * 1. Configure your memory controller to add such delays if it can (the best)
21  * 2. Implement platform-specific delay function possibly
22  *    combined with configuring the memory controller; see
23  *    include/linux/usb_isp1362.h for more info.
24  * 3. Use ndelay (easiest, poorest).
25  *
26  * Use the corresponding macros USE_PLATFORM_DELAY and USE_NDELAY in the
27  * platform specific section of isp1362.h to select the appropriate variant.
28  *
29  * Also note that according to the Philips "ISP1362 Errata" document
30  * Rev 1.00 from 27 May data corruption may occur when the #WR signal
31  * is reasserted (even with #CS deasserted) within 132ns after a
32  * write cycle to any controller register. If the hardware doesn't
33  * implement the recommended fix (gating the #WR with #CS) software
34  * must ensure that no further write cycle (not necessarily to the chip!)
35  * is issued by the CPU within this interval.
36 
37  * For PXA25x this can be ensured by using VLIO with the maximum
38  * recovery time (MSCx = 0x7f8c) with a memory clock of 99.53 MHz.
39  */
40 
41 #undef ISP1362_DEBUG
42 
43 /*
44  * The PXA255 UDC apparently doesn't handle GET_STATUS, GET_CONFIG and
45  * GET_INTERFACE requests correctly when the SETUP and DATA stages of the
46  * requests are carried out in separate frames. This will delay any SETUP
47  * packets until the start of the next frame so that this situation is
48  * unlikely to occur (and makes usbtest happy running with a PXA255 target
49  * device).
50  */
51 #undef BUGGY_PXA2XX_UDC_USBTEST
52 
53 #undef PTD_TRACE
54 #undef URB_TRACE
55 #undef VERBOSE
56 #undef REGISTERS
57 
58 /* This enables a memory test on the ISP1362 chip memory to make sure the
59  * chip access timing is correct.
60  */
61 #undef CHIP_BUFFER_TEST
62 
63 #include <linux/module.h>
64 #include <linux/moduleparam.h>
65 #include <linux/kernel.h>
66 #include <linux/delay.h>
67 #include <linux/ioport.h>
68 #include <linux/sched.h>
69 #include <linux/slab.h>
70 #include <linux/errno.h>
71 #include <linux/list.h>
72 #include <linux/interrupt.h>
73 #include <linux/usb.h>
74 #include <linux/usb/isp1362.h>
75 #include <linux/usb/hcd.h>
76 #include <linux/platform_device.h>
77 #include <linux/pm.h>
78 #include <linux/io.h>
79 #include <linux/bitmap.h>
80 #include <linux/prefetch.h>
81 #include <linux/debugfs.h>
82 #include <linux/seq_file.h>
83 
84 #include <asm/irq.h>
85 #include <asm/byteorder.h>
86 #include <asm/unaligned.h>
87 
88 static int dbg_level;
89 #ifdef ISP1362_DEBUG
90 module_param(dbg_level, int, 0644);
91 #else
92 module_param(dbg_level, int, 0);
93 #endif
94 
95 #include "../core/usb.h"
96 #include "isp1362.h"
97 
98 
99 #define DRIVER_VERSION	"2005-04-04"
100 #define DRIVER_DESC	"ISP1362 USB Host Controller Driver"
101 
102 MODULE_DESCRIPTION(DRIVER_DESC);
103 MODULE_LICENSE("GPL");
104 
105 static const char hcd_name[] = "isp1362-hcd";
106 
107 static void isp1362_hc_stop(struct usb_hcd *hcd);
108 static int isp1362_hc_start(struct usb_hcd *hcd);
109 
110 /*-------------------------------------------------------------------------*/
111 
112 /*
113  * When called from the interrupthandler only isp1362_hcd->irqenb is modified,
114  * since the interrupt handler will write isp1362_hcd->irqenb to HCuPINT upon
115  * completion.
116  * We don't need a 'disable' counterpart, since interrupts will be disabled
117  * only by the interrupt handler.
118  */
119 static inline void isp1362_enable_int(struct isp1362_hcd *isp1362_hcd, u16 mask)
120 {
121 	if ((isp1362_hcd->irqenb | mask) == isp1362_hcd->irqenb)
122 		return;
123 	if (mask & ~isp1362_hcd->irqenb)
124 		isp1362_write_reg16(isp1362_hcd, HCuPINT, mask & ~isp1362_hcd->irqenb);
125 	isp1362_hcd->irqenb |= mask;
126 	if (isp1362_hcd->irq_active)
127 		return;
128 	isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
129 }
130 
131 /*-------------------------------------------------------------------------*/
132 
133 static inline struct isp1362_ep_queue *get_ptd_queue(struct isp1362_hcd *isp1362_hcd,
134 						     u16 offset)
135 {
136 	struct isp1362_ep_queue *epq = NULL;
137 
138 	if (offset < isp1362_hcd->istl_queue[1].buf_start)
139 		epq = &isp1362_hcd->istl_queue[0];
140 	else if (offset < isp1362_hcd->intl_queue.buf_start)
141 		epq = &isp1362_hcd->istl_queue[1];
142 	else if (offset < isp1362_hcd->atl_queue.buf_start)
143 		epq = &isp1362_hcd->intl_queue;
144 	else if (offset < isp1362_hcd->atl_queue.buf_start +
145 		   isp1362_hcd->atl_queue.buf_size)
146 		epq = &isp1362_hcd->atl_queue;
147 
148 	if (epq)
149 		DBG(1, "%s: PTD $%04x is on %s queue\n", __func__, offset, epq->name);
150 	else
151 		pr_warn("%s: invalid PTD $%04x\n", __func__, offset);
152 
153 	return epq;
154 }
155 
156 static inline int get_ptd_offset(struct isp1362_ep_queue *epq, u8 index)
157 {
158 	int offset;
159 
160 	if (index * epq->blk_size > epq->buf_size) {
161 		pr_warn("%s: Bad %s index %d(%d)\n",
162 			__func__, epq->name, index,
163 			epq->buf_size / epq->blk_size);
164 		return -EINVAL;
165 	}
166 	offset = epq->buf_start + index * epq->blk_size;
167 	DBG(3, "%s: %s PTD[%02x] # %04x\n", __func__, epq->name, index, offset);
168 
169 	return offset;
170 }
171 
172 /*-------------------------------------------------------------------------*/
173 
174 static inline u16 max_transfer_size(struct isp1362_ep_queue *epq, size_t size,
175 				    int mps)
176 {
177 	u16 xfer_size = min_t(size_t, MAX_XFER_SIZE, size);
178 
179 	xfer_size = min_t(size_t, xfer_size, epq->buf_avail * epq->blk_size - PTD_HEADER_SIZE);
180 	if (xfer_size < size && xfer_size % mps)
181 		xfer_size -= xfer_size % mps;
182 
183 	return xfer_size;
184 }
185 
186 static int claim_ptd_buffers(struct isp1362_ep_queue *epq,
187 			     struct isp1362_ep *ep, u16 len)
188 {
189 	int ptd_offset = -EINVAL;
190 	int num_ptds = ((len + PTD_HEADER_SIZE - 1) / epq->blk_size) + 1;
191 	int found;
192 
193 	BUG_ON(len > epq->buf_size);
194 
195 	if (!epq->buf_avail)
196 		return -ENOMEM;
197 
198 	if (ep->num_ptds)
199 		pr_err("%s: %s len %d/%d num_ptds %d buf_map %08lx skip_map %08lx\n", __func__,
200 		    epq->name, len, epq->blk_size, num_ptds, epq->buf_map, epq->skip_map);
201 	BUG_ON(ep->num_ptds != 0);
202 
203 	found = bitmap_find_next_zero_area(&epq->buf_map, epq->buf_count, 0,
204 						num_ptds, 0);
205 	if (found >= epq->buf_count)
206 		return -EOVERFLOW;
207 
208 	DBG(1, "%s: Found %d PTDs[%d] for %d/%d byte\n", __func__,
209 	    num_ptds, found, len, (int)(epq->blk_size - PTD_HEADER_SIZE));
210 	ptd_offset = get_ptd_offset(epq, found);
211 	WARN_ON(ptd_offset < 0);
212 	ep->ptd_offset = ptd_offset;
213 	ep->num_ptds += num_ptds;
214 	epq->buf_avail -= num_ptds;
215 	BUG_ON(epq->buf_avail > epq->buf_count);
216 	ep->ptd_index = found;
217 	bitmap_set(&epq->buf_map, found, num_ptds);
218 	DBG(1, "%s: Done %s PTD[%d] $%04x, avail %d count %d claimed %d %08lx:%08lx\n",
219 	    __func__, epq->name, ep->ptd_index, ep->ptd_offset,
220 	    epq->buf_avail, epq->buf_count, num_ptds, epq->buf_map, epq->skip_map);
221 
222 	return found;
223 }
224 
225 static inline void release_ptd_buffers(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
226 {
227 	int last = ep->ptd_index + ep->num_ptds;
228 
229 	if (last > epq->buf_count)
230 		pr_err("%s: ep %p req %d len %d %s PTD[%d] $%04x num_ptds %d buf_count %d buf_avail %d buf_map %08lx skip_map %08lx\n",
231 		    __func__, ep, ep->num_req, ep->length, epq->name, ep->ptd_index,
232 		    ep->ptd_offset, ep->num_ptds, epq->buf_count, epq->buf_avail,
233 		    epq->buf_map, epq->skip_map);
234 	BUG_ON(last > epq->buf_count);
235 
236 	bitmap_clear(&epq->buf_map, ep->ptd_index, ep->num_ptds);
237 	bitmap_set(&epq->skip_map, ep->ptd_index, ep->num_ptds);
238 	epq->buf_avail += ep->num_ptds;
239 	epq->ptd_count--;
240 
241 	BUG_ON(epq->buf_avail > epq->buf_count);
242 	BUG_ON(epq->ptd_count > epq->buf_count);
243 
244 	DBG(1, "%s: Done %s PTDs $%04x released %d avail %d count %d\n",
245 	    __func__, epq->name,
246 	    ep->ptd_offset, ep->num_ptds, epq->buf_avail, epq->buf_count);
247 	DBG(1, "%s: buf_map %08lx skip_map %08lx\n", __func__,
248 	    epq->buf_map, epq->skip_map);
249 
250 	ep->num_ptds = 0;
251 	ep->ptd_offset = -EINVAL;
252 	ep->ptd_index = -EINVAL;
253 }
254 
255 /*-------------------------------------------------------------------------*/
256 
257 /*
258   Set up PTD's.
259 */
260 static void prepare_ptd(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
261 			struct isp1362_ep *ep, struct isp1362_ep_queue *epq,
262 			u16 fno)
263 {
264 	struct ptd *ptd;
265 	int toggle;
266 	int dir;
267 	u16 len;
268 	size_t buf_len = urb->transfer_buffer_length - urb->actual_length;
269 
270 	DBG(3, "%s: %s ep %p\n", __func__, epq->name, ep);
271 
272 	ptd = &ep->ptd;
273 
274 	ep->data = (unsigned char *)urb->transfer_buffer + urb->actual_length;
275 
276 	switch (ep->nextpid) {
277 	case USB_PID_IN:
278 		toggle = usb_gettoggle(urb->dev, ep->epnum, 0);
279 		dir = PTD_DIR_IN;
280 		if (usb_pipecontrol(urb->pipe)) {
281 			len = min_t(size_t, ep->maxpacket, buf_len);
282 		} else if (usb_pipeisoc(urb->pipe)) {
283 			len = min_t(size_t, urb->iso_frame_desc[fno].length, MAX_XFER_SIZE);
284 			ep->data = urb->transfer_buffer + urb->iso_frame_desc[fno].offset;
285 		} else
286 			len = max_transfer_size(epq, buf_len, ep->maxpacket);
287 		DBG(1, "%s: IN    len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
288 		    (int)buf_len);
289 		break;
290 	case USB_PID_OUT:
291 		toggle = usb_gettoggle(urb->dev, ep->epnum, 1);
292 		dir = PTD_DIR_OUT;
293 		if (usb_pipecontrol(urb->pipe))
294 			len = min_t(size_t, ep->maxpacket, buf_len);
295 		else if (usb_pipeisoc(urb->pipe))
296 			len = min_t(size_t, urb->iso_frame_desc[0].length, MAX_XFER_SIZE);
297 		else
298 			len = max_transfer_size(epq, buf_len, ep->maxpacket);
299 		if (len == 0)
300 			pr_info("%s: Sending ZERO packet: %d\n", __func__,
301 			     urb->transfer_flags & URB_ZERO_PACKET);
302 		DBG(1, "%s: OUT   len %d/%d/%d from URB\n", __func__, len, ep->maxpacket,
303 		    (int)buf_len);
304 		break;
305 	case USB_PID_SETUP:
306 		toggle = 0;
307 		dir = PTD_DIR_SETUP;
308 		len = sizeof(struct usb_ctrlrequest);
309 		DBG(1, "%s: SETUP len %d\n", __func__, len);
310 		ep->data = urb->setup_packet;
311 		break;
312 	case USB_PID_ACK:
313 		toggle = 1;
314 		len = 0;
315 		dir = (urb->transfer_buffer_length && usb_pipein(urb->pipe)) ?
316 			PTD_DIR_OUT : PTD_DIR_IN;
317 		DBG(1, "%s: ACK   len %d\n", __func__, len);
318 		break;
319 	default:
320 		toggle = dir = len = 0;
321 		pr_err("%s@%d: ep->nextpid %02x\n", __func__, __LINE__, ep->nextpid);
322 		BUG_ON(1);
323 	}
324 
325 	ep->length = len;
326 	if (!len)
327 		ep->data = NULL;
328 
329 	ptd->count = PTD_CC_MSK | PTD_ACTIVE_MSK | PTD_TOGGLE(toggle);
330 	ptd->mps = PTD_MPS(ep->maxpacket) | PTD_SPD(urb->dev->speed == USB_SPEED_LOW) |
331 		PTD_EP(ep->epnum);
332 	ptd->len = PTD_LEN(len) | PTD_DIR(dir);
333 	ptd->faddr = PTD_FA(usb_pipedevice(urb->pipe));
334 
335 	if (usb_pipeint(urb->pipe)) {
336 		ptd->faddr |= PTD_SF_INT(ep->branch);
337 		ptd->faddr |= PTD_PR(ep->interval ? __ffs(ep->interval) : 0);
338 	}
339 	if (usb_pipeisoc(urb->pipe))
340 		ptd->faddr |= PTD_SF_ISO(fno);
341 
342 	DBG(1, "%s: Finished\n", __func__);
343 }
344 
345 static void isp1362_write_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
346 			      struct isp1362_ep_queue *epq)
347 {
348 	struct ptd *ptd = &ep->ptd;
349 	int len = PTD_GET_DIR(ptd) == PTD_DIR_IN ? 0 : ep->length;
350 
351 	prefetch(ptd);
352 	isp1362_write_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
353 	if (len)
354 		isp1362_write_buffer(isp1362_hcd, ep->data,
355 				     ep->ptd_offset + PTD_HEADER_SIZE, len);
356 
357 	dump_ptd(ptd);
358 	dump_ptd_out_data(ptd, ep->data);
359 }
360 
361 static void isp1362_read_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
362 			     struct isp1362_ep_queue *epq)
363 {
364 	struct ptd *ptd = &ep->ptd;
365 	int act_len;
366 
367 	WARN_ON(list_empty(&ep->active));
368 	BUG_ON(ep->ptd_offset < 0);
369 
370 	list_del_init(&ep->active);
371 	DBG(1, "%s: ep %p removed from active list %p\n", __func__, ep, &epq->active);
372 
373 	prefetchw(ptd);
374 	isp1362_read_buffer(isp1362_hcd, ptd, ep->ptd_offset, PTD_HEADER_SIZE);
375 	dump_ptd(ptd);
376 	act_len = PTD_GET_COUNT(ptd);
377 	if (PTD_GET_DIR(ptd) != PTD_DIR_IN || act_len == 0)
378 		return;
379 	if (act_len > ep->length)
380 		pr_err("%s: ep %p PTD $%04x act_len %d ep->length %d\n", __func__, ep,
381 			 ep->ptd_offset, act_len, ep->length);
382 	BUG_ON(act_len > ep->length);
383 	/* Only transfer the amount of data that has actually been overwritten
384 	 * in the chip buffer. We don't want any data that doesn't belong to the
385 	 * transfer to leak out of the chip to the callers transfer buffer!
386 	 */
387 	prefetchw(ep->data);
388 	isp1362_read_buffer(isp1362_hcd, ep->data,
389 			    ep->ptd_offset + PTD_HEADER_SIZE, act_len);
390 	dump_ptd_in_data(ptd, ep->data);
391 }
392 
393 /*
394  * INT PTDs will stay in the chip until data is available.
395  * This function will remove a PTD from the chip when the URB is dequeued.
396  * Must be called with the spinlock held and IRQs disabled
397  */
398 static void remove_ptd(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
399 
400 {
401 	int index;
402 	struct isp1362_ep_queue *epq;
403 
404 	DBG(1, "%s: ep %p PTD[%d] $%04x\n", __func__, ep, ep->ptd_index, ep->ptd_offset);
405 	BUG_ON(ep->ptd_offset < 0);
406 
407 	epq = get_ptd_queue(isp1362_hcd, ep->ptd_offset);
408 	BUG_ON(!epq);
409 
410 	/* put ep in remove_list for cleanup */
411 	WARN_ON(!list_empty(&ep->remove_list));
412 	list_add_tail(&ep->remove_list, &isp1362_hcd->remove_list);
413 	/* let SOF interrupt handle the cleanup */
414 	isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
415 
416 	index = ep->ptd_index;
417 	if (index < 0)
418 		/* ISO queues don't have SKIP registers */
419 		return;
420 
421 	DBG(1, "%s: Disabling PTD[%02x] $%04x %08lx|%08x\n", __func__,
422 	    index, ep->ptd_offset, epq->skip_map, 1 << index);
423 
424 	/* prevent further processing of PTD (will be effective after next SOF) */
425 	epq->skip_map |= 1 << index;
426 	if (epq == &isp1362_hcd->atl_queue) {
427 		DBG(2, "%s: ATLSKIP = %08x -> %08lx\n", __func__,
428 		    isp1362_read_reg32(isp1362_hcd, HCATLSKIP), epq->skip_map);
429 		isp1362_write_reg32(isp1362_hcd, HCATLSKIP, epq->skip_map);
430 		if (~epq->skip_map == 0)
431 			isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
432 	} else if (epq == &isp1362_hcd->intl_queue) {
433 		DBG(2, "%s: INTLSKIP = %08x -> %08lx\n", __func__,
434 		    isp1362_read_reg32(isp1362_hcd, HCINTLSKIP), epq->skip_map);
435 		isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, epq->skip_map);
436 		if (~epq->skip_map == 0)
437 			isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
438 	}
439 }
440 
441 /*
442   Take done or failed requests out of schedule. Give back
443   processed urbs.
444 */
445 static void finish_request(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep,
446 			   struct urb *urb, int status)
447      __releases(isp1362_hcd->lock)
448      __acquires(isp1362_hcd->lock)
449 {
450 	urb->hcpriv = NULL;
451 	ep->error_count = 0;
452 
453 	if (usb_pipecontrol(urb->pipe))
454 		ep->nextpid = USB_PID_SETUP;
455 
456 	URB_DBG("%s: req %d FA %d ep%d%s %s: len %d/%d %s stat %d\n", __func__,
457 		ep->num_req, usb_pipedevice(urb->pipe),
458 		usb_pipeendpoint(urb->pipe),
459 		!usb_pipein(urb->pipe) ? "out" : "in",
460 		usb_pipecontrol(urb->pipe) ? "ctrl" :
461 			usb_pipeint(urb->pipe) ? "int" :
462 			usb_pipebulk(urb->pipe) ? "bulk" :
463 			"iso",
464 		urb->actual_length, urb->transfer_buffer_length,
465 		!(urb->transfer_flags & URB_SHORT_NOT_OK) ?
466 		"short_ok" : "", urb->status);
467 
468 
469 	usb_hcd_unlink_urb_from_ep(isp1362_hcd_to_hcd(isp1362_hcd), urb);
470 	spin_unlock(&isp1362_hcd->lock);
471 	usb_hcd_giveback_urb(isp1362_hcd_to_hcd(isp1362_hcd), urb, status);
472 	spin_lock(&isp1362_hcd->lock);
473 
474 	/* take idle endpoints out of the schedule right away */
475 	if (!list_empty(&ep->hep->urb_list))
476 		return;
477 
478 	/* async deschedule */
479 	if (!list_empty(&ep->schedule)) {
480 		list_del_init(&ep->schedule);
481 		return;
482 	}
483 
484 
485 	if (ep->interval) {
486 		/* periodic deschedule */
487 		DBG(1, "deschedule qh%d/%p branch %d load %d bandwidth %d -> %d\n", ep->interval,
488 		    ep, ep->branch, ep->load,
489 		    isp1362_hcd->load[ep->branch],
490 		    isp1362_hcd->load[ep->branch] - ep->load);
491 		isp1362_hcd->load[ep->branch] -= ep->load;
492 		ep->branch = PERIODIC_SIZE;
493 	}
494 }
495 
496 /*
497  * Analyze transfer results, handle partial transfers and errors
498 */
499 static void postproc_ep(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep *ep)
500 {
501 	struct urb *urb = get_urb(ep);
502 	struct usb_device *udev;
503 	struct ptd *ptd;
504 	int short_ok;
505 	u16 len;
506 	int urbstat = -EINPROGRESS;
507 	u8 cc;
508 
509 	DBG(2, "%s: ep %p req %d\n", __func__, ep, ep->num_req);
510 
511 	udev = urb->dev;
512 	ptd = &ep->ptd;
513 	cc = PTD_GET_CC(ptd);
514 	if (cc == PTD_NOTACCESSED) {
515 		pr_err("%s: req %d PTD %p Untouched by ISP1362\n", __func__,
516 		    ep->num_req, ptd);
517 		cc = PTD_DEVNOTRESP;
518 	}
519 
520 	short_ok = !(urb->transfer_flags & URB_SHORT_NOT_OK);
521 	len = urb->transfer_buffer_length - urb->actual_length;
522 
523 	/* Data underrun is special. For allowed underrun
524 	   we clear the error and continue as normal. For
525 	   forbidden underrun we finish the DATA stage
526 	   immediately while for control transfer,
527 	   we do a STATUS stage.
528 	*/
529 	if (cc == PTD_DATAUNDERRUN) {
530 		if (short_ok) {
531 			DBG(1, "%s: req %d Allowed data underrun short_%sok %d/%d/%d byte\n",
532 			    __func__, ep->num_req, short_ok ? "" : "not_",
533 			    PTD_GET_COUNT(ptd), ep->maxpacket, len);
534 			cc = PTD_CC_NOERROR;
535 			urbstat = 0;
536 		} else {
537 			DBG(1, "%s: req %d Data Underrun %s nextpid %02x short_%sok %d/%d/%d byte\n",
538 			    __func__, ep->num_req,
539 			    usb_pipein(urb->pipe) ? "IN" : "OUT", ep->nextpid,
540 			    short_ok ? "" : "not_",
541 			    PTD_GET_COUNT(ptd), ep->maxpacket, len);
542 			/* save the data underrun error code for later and
543 			 * proceed with the status stage
544 			 */
545 			urb->actual_length += PTD_GET_COUNT(ptd);
546 			if (usb_pipecontrol(urb->pipe)) {
547 				ep->nextpid = USB_PID_ACK;
548 				BUG_ON(urb->actual_length > urb->transfer_buffer_length);
549 
550 				if (urb->status == -EINPROGRESS)
551 					urb->status = cc_to_error[PTD_DATAUNDERRUN];
552 			} else {
553 				usb_settoggle(udev, ep->epnum, ep->nextpid == USB_PID_OUT,
554 					      PTD_GET_TOGGLE(ptd));
555 				urbstat = cc_to_error[PTD_DATAUNDERRUN];
556 			}
557 			goto out;
558 		}
559 	}
560 
561 	if (cc != PTD_CC_NOERROR) {
562 		if (++ep->error_count >= 3 || cc == PTD_CC_STALL || cc == PTD_DATAOVERRUN) {
563 			urbstat = cc_to_error[cc];
564 			DBG(1, "%s: req %d nextpid %02x, status %d, error %d, error_count %d\n",
565 			    __func__, ep->num_req, ep->nextpid, urbstat, cc,
566 			    ep->error_count);
567 		}
568 		goto out;
569 	}
570 
571 	switch (ep->nextpid) {
572 	case USB_PID_OUT:
573 		if (PTD_GET_COUNT(ptd) != ep->length)
574 			pr_err("%s: count=%d len=%d\n", __func__,
575 			   PTD_GET_COUNT(ptd), ep->length);
576 		BUG_ON(PTD_GET_COUNT(ptd) != ep->length);
577 		urb->actual_length += ep->length;
578 		BUG_ON(urb->actual_length > urb->transfer_buffer_length);
579 		usb_settoggle(udev, ep->epnum, 1, PTD_GET_TOGGLE(ptd));
580 		if (urb->actual_length == urb->transfer_buffer_length) {
581 			DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
582 			    ep->num_req, len, ep->maxpacket, urbstat);
583 			if (usb_pipecontrol(urb->pipe)) {
584 				DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
585 				    ep->num_req,
586 				    usb_pipein(urb->pipe) ? "IN" : "OUT");
587 				ep->nextpid = USB_PID_ACK;
588 			} else {
589 				if (len % ep->maxpacket ||
590 				    !(urb->transfer_flags & URB_ZERO_PACKET)) {
591 					urbstat = 0;
592 					DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
593 					    __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
594 					    urbstat, len, ep->maxpacket, urb->actual_length);
595 				}
596 			}
597 		}
598 		break;
599 	case USB_PID_IN:
600 		len = PTD_GET_COUNT(ptd);
601 		BUG_ON(len > ep->length);
602 		urb->actual_length += len;
603 		BUG_ON(urb->actual_length > urb->transfer_buffer_length);
604 		usb_settoggle(udev, ep->epnum, 0, PTD_GET_TOGGLE(ptd));
605 		/* if transfer completed or (allowed) data underrun */
606 		if ((urb->transfer_buffer_length == urb->actual_length) ||
607 		    len % ep->maxpacket) {
608 			DBG(3, "%s: req %d xfer complete %d/%d status %d -> 0\n", __func__,
609 			    ep->num_req, len, ep->maxpacket, urbstat);
610 			if (usb_pipecontrol(urb->pipe)) {
611 				DBG(3, "%s: req %d %s Wait for ACK\n", __func__,
612 				    ep->num_req,
613 				    usb_pipein(urb->pipe) ? "IN" : "OUT");
614 				ep->nextpid = USB_PID_ACK;
615 			} else {
616 				urbstat = 0;
617 				DBG(3, "%s: req %d URB %s status %d count %d/%d/%d\n",
618 				    __func__, ep->num_req, usb_pipein(urb->pipe) ? "IN" : "OUT",
619 				    urbstat, len, ep->maxpacket, urb->actual_length);
620 			}
621 		}
622 		break;
623 	case USB_PID_SETUP:
624 		if (urb->transfer_buffer_length == urb->actual_length) {
625 			ep->nextpid = USB_PID_ACK;
626 		} else if (usb_pipeout(urb->pipe)) {
627 			usb_settoggle(udev, 0, 1, 1);
628 			ep->nextpid = USB_PID_OUT;
629 		} else {
630 			usb_settoggle(udev, 0, 0, 1);
631 			ep->nextpid = USB_PID_IN;
632 		}
633 		break;
634 	case USB_PID_ACK:
635 		DBG(3, "%s: req %d got ACK %d -> 0\n", __func__, ep->num_req,
636 		    urbstat);
637 		WARN_ON(urbstat != -EINPROGRESS);
638 		urbstat = 0;
639 		ep->nextpid = 0;
640 		break;
641 	default:
642 		BUG_ON(1);
643 	}
644 
645  out:
646 	if (urbstat != -EINPROGRESS) {
647 		DBG(2, "%s: Finishing ep %p req %d urb %p status %d\n", __func__,
648 		    ep, ep->num_req, urb, urbstat);
649 		finish_request(isp1362_hcd, ep, urb, urbstat);
650 	}
651 }
652 
653 static void finish_unlinks(struct isp1362_hcd *isp1362_hcd)
654 {
655 	struct isp1362_ep *ep;
656 	struct isp1362_ep *tmp;
657 
658 	list_for_each_entry_safe(ep, tmp, &isp1362_hcd->remove_list, remove_list) {
659 		struct isp1362_ep_queue *epq =
660 			get_ptd_queue(isp1362_hcd, ep->ptd_offset);
661 		int index = ep->ptd_index;
662 
663 		BUG_ON(epq == NULL);
664 		if (index >= 0) {
665 			DBG(1, "%s: remove PTD[%d] $%04x\n", __func__, index, ep->ptd_offset);
666 			BUG_ON(ep->num_ptds == 0);
667 			release_ptd_buffers(epq, ep);
668 		}
669 		if (!list_empty(&ep->hep->urb_list)) {
670 			struct urb *urb = get_urb(ep);
671 
672 			DBG(1, "%s: Finishing req %d ep %p from remove_list\n", __func__,
673 			    ep->num_req, ep);
674 			finish_request(isp1362_hcd, ep, urb, -ESHUTDOWN);
675 		}
676 		WARN_ON(list_empty(&ep->active));
677 		if (!list_empty(&ep->active)) {
678 			list_del_init(&ep->active);
679 			DBG(1, "%s: ep %p removed from active list\n", __func__, ep);
680 		}
681 		list_del_init(&ep->remove_list);
682 		DBG(1, "%s: ep %p removed from remove_list\n", __func__, ep);
683 	}
684 	DBG(1, "%s: Done\n", __func__);
685 }
686 
687 static inline void enable_atl_transfers(struct isp1362_hcd *isp1362_hcd, int count)
688 {
689 	if (count > 0) {
690 		if (count < isp1362_hcd->atl_queue.ptd_count)
691 			isp1362_write_reg16(isp1362_hcd, HCATLDTC, count);
692 		isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
693 		isp1362_write_reg32(isp1362_hcd, HCATLSKIP, isp1362_hcd->atl_queue.skip_map);
694 		isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
695 	} else
696 		isp1362_enable_int(isp1362_hcd, HCuPINT_SOF);
697 }
698 
699 static inline void enable_intl_transfers(struct isp1362_hcd *isp1362_hcd)
700 {
701 	isp1362_enable_int(isp1362_hcd, HCuPINT_INTL);
702 	isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
703 	isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, isp1362_hcd->intl_queue.skip_map);
704 }
705 
706 static inline void enable_istl_transfers(struct isp1362_hcd *isp1362_hcd, int flip)
707 {
708 	isp1362_enable_int(isp1362_hcd, flip ? HCuPINT_ISTL1 : HCuPINT_ISTL0);
709 	isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, flip ?
710 			   HCBUFSTAT_ISTL1_FULL : HCBUFSTAT_ISTL0_FULL);
711 }
712 
713 static int submit_req(struct isp1362_hcd *isp1362_hcd, struct urb *urb,
714 		      struct isp1362_ep *ep, struct isp1362_ep_queue *epq)
715 {
716 	int index;
717 
718 	prepare_ptd(isp1362_hcd, urb, ep, epq, 0);
719 	index = claim_ptd_buffers(epq, ep, ep->length);
720 	if (index == -ENOMEM) {
721 		DBG(1, "%s: req %d No free %s PTD available: %d, %08lx:%08lx\n", __func__,
722 		    ep->num_req, epq->name, ep->num_ptds, epq->buf_map, epq->skip_map);
723 		return index;
724 	} else if (index == -EOVERFLOW) {
725 		DBG(1, "%s: req %d Not enough space for %d byte %s PTD %d %08lx:%08lx\n",
726 		    __func__, ep->num_req, ep->length, epq->name, ep->num_ptds,
727 		    epq->buf_map, epq->skip_map);
728 		return index;
729 	} else
730 		BUG_ON(index < 0);
731 	list_add_tail(&ep->active, &epq->active);
732 	DBG(1, "%s: ep %p req %d len %d added to active list %p\n", __func__,
733 	    ep, ep->num_req, ep->length, &epq->active);
734 	DBG(1, "%s: Submitting %s PTD $%04x for ep %p req %d\n", __func__, epq->name,
735 	    ep->ptd_offset, ep, ep->num_req);
736 	isp1362_write_ptd(isp1362_hcd, ep, epq);
737 	__clear_bit(ep->ptd_index, &epq->skip_map);
738 
739 	return 0;
740 }
741 
742 static void start_atl_transfers(struct isp1362_hcd *isp1362_hcd)
743 {
744 	int ptd_count = 0;
745 	struct isp1362_ep_queue *epq = &isp1362_hcd->atl_queue;
746 	struct isp1362_ep *ep;
747 	int defer = 0;
748 
749 	if (atomic_read(&epq->finishing)) {
750 		DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
751 		return;
752 	}
753 
754 	list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
755 		struct urb *urb = get_urb(ep);
756 		int ret;
757 
758 		if (!list_empty(&ep->active)) {
759 			DBG(2, "%s: Skipping active %s ep %p\n", __func__, epq->name, ep);
760 			continue;
761 		}
762 
763 		DBG(1, "%s: Processing %s ep %p req %d\n", __func__, epq->name,
764 		    ep, ep->num_req);
765 
766 		ret = submit_req(isp1362_hcd, urb, ep, epq);
767 		if (ret == -ENOMEM) {
768 			defer = 1;
769 			break;
770 		} else if (ret == -EOVERFLOW) {
771 			defer = 1;
772 			continue;
773 		}
774 #ifdef BUGGY_PXA2XX_UDC_USBTEST
775 		defer = ep->nextpid == USB_PID_SETUP;
776 #endif
777 		ptd_count++;
778 	}
779 
780 	/* Avoid starving of endpoints */
781 	if (isp1362_hcd->async.next != isp1362_hcd->async.prev) {
782 		DBG(2, "%s: Cycling ASYNC schedule %d\n", __func__, ptd_count);
783 		list_move(&isp1362_hcd->async, isp1362_hcd->async.next);
784 	}
785 	if (ptd_count || defer)
786 		enable_atl_transfers(isp1362_hcd, defer ? 0 : ptd_count);
787 
788 	epq->ptd_count += ptd_count;
789 	if (epq->ptd_count > epq->stat_maxptds) {
790 		epq->stat_maxptds = epq->ptd_count;
791 		DBG(0, "%s: max_ptds: %d\n", __func__, epq->stat_maxptds);
792 	}
793 }
794 
795 static void start_intl_transfers(struct isp1362_hcd *isp1362_hcd)
796 {
797 	int ptd_count = 0;
798 	struct isp1362_ep_queue *epq = &isp1362_hcd->intl_queue;
799 	struct isp1362_ep *ep;
800 
801 	if (atomic_read(&epq->finishing)) {
802 		DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
803 		return;
804 	}
805 
806 	list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
807 		struct urb *urb = get_urb(ep);
808 		int ret;
809 
810 		if (!list_empty(&ep->active)) {
811 			DBG(1, "%s: Skipping active %s ep %p\n", __func__,
812 			    epq->name, ep);
813 			continue;
814 		}
815 
816 		DBG(1, "%s: Processing %s ep %p req %d\n", __func__,
817 		    epq->name, ep, ep->num_req);
818 		ret = submit_req(isp1362_hcd, urb, ep, epq);
819 		if (ret == -ENOMEM)
820 			break;
821 		else if (ret == -EOVERFLOW)
822 			continue;
823 		ptd_count++;
824 	}
825 
826 	if (ptd_count) {
827 		static int last_count;
828 
829 		if (ptd_count != last_count) {
830 			DBG(0, "%s: ptd_count: %d\n", __func__, ptd_count);
831 			last_count = ptd_count;
832 		}
833 		enable_intl_transfers(isp1362_hcd);
834 	}
835 
836 	epq->ptd_count += ptd_count;
837 	if (epq->ptd_count > epq->stat_maxptds)
838 		epq->stat_maxptds = epq->ptd_count;
839 }
840 
841 static inline int next_ptd(struct isp1362_ep_queue *epq, struct isp1362_ep *ep)
842 {
843 	u16 ptd_offset = ep->ptd_offset;
844 	int num_ptds = (ep->length + PTD_HEADER_SIZE + (epq->blk_size - 1)) / epq->blk_size;
845 
846 	DBG(2, "%s: PTD offset $%04x + %04x => %d * %04x -> $%04x\n", __func__, ptd_offset,
847 	    ep->length, num_ptds, epq->blk_size, ptd_offset + num_ptds * epq->blk_size);
848 
849 	ptd_offset += num_ptds * epq->blk_size;
850 	if (ptd_offset < epq->buf_start + epq->buf_size)
851 		return ptd_offset;
852 	else
853 		return -ENOMEM;
854 }
855 
856 static void start_iso_transfers(struct isp1362_hcd *isp1362_hcd)
857 {
858 	int ptd_count = 0;
859 	int flip = isp1362_hcd->istl_flip;
860 	struct isp1362_ep_queue *epq;
861 	int ptd_offset;
862 	struct isp1362_ep *ep;
863 	struct isp1362_ep *tmp;
864 	u16 fno = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
865 
866  fill2:
867 	epq = &isp1362_hcd->istl_queue[flip];
868 	if (atomic_read(&epq->finishing)) {
869 		DBG(1, "%s: finish_transfers is active for %s\n", __func__, epq->name);
870 		return;
871 	}
872 
873 	if (!list_empty(&epq->active))
874 		return;
875 
876 	ptd_offset = epq->buf_start;
877 	list_for_each_entry_safe(ep, tmp, &isp1362_hcd->isoc, schedule) {
878 		struct urb *urb = get_urb(ep);
879 		s16 diff = fno - (u16)urb->start_frame;
880 
881 		DBG(1, "%s: Processing %s ep %p\n", __func__, epq->name, ep);
882 
883 		if (diff > urb->number_of_packets) {
884 			/* time frame for this URB has elapsed */
885 			finish_request(isp1362_hcd, ep, urb, -EOVERFLOW);
886 			continue;
887 		} else if (diff < -1) {
888 			/* URB is not due in this frame or the next one.
889 			 * Comparing with '-1' instead of '0' accounts for double
890 			 * buffering in the ISP1362 which enables us to queue the PTD
891 			 * one frame ahead of time
892 			 */
893 		} else if (diff == -1) {
894 			/* submit PTD's that are due in the next frame */
895 			prepare_ptd(isp1362_hcd, urb, ep, epq, fno);
896 			if (ptd_offset + PTD_HEADER_SIZE + ep->length >
897 			    epq->buf_start + epq->buf_size) {
898 				pr_err("%s: Not enough ISO buffer space for %d byte PTD\n",
899 				    __func__, ep->length);
900 				continue;
901 			}
902 			ep->ptd_offset = ptd_offset;
903 			list_add_tail(&ep->active, &epq->active);
904 
905 			ptd_offset = next_ptd(epq, ep);
906 			if (ptd_offset < 0) {
907 				pr_warn("%s: req %d No more %s PTD buffers available\n",
908 					__func__, ep->num_req, epq->name);
909 				break;
910 			}
911 		}
912 	}
913 	list_for_each_entry(ep, &epq->active, active) {
914 		if (epq->active.next == &ep->active)
915 			ep->ptd.mps |= PTD_LAST_MSK;
916 		isp1362_write_ptd(isp1362_hcd, ep, epq);
917 		ptd_count++;
918 	}
919 
920 	if (ptd_count)
921 		enable_istl_transfers(isp1362_hcd, flip);
922 
923 	epq->ptd_count += ptd_count;
924 	if (epq->ptd_count > epq->stat_maxptds)
925 		epq->stat_maxptds = epq->ptd_count;
926 
927 	/* check, whether the second ISTL buffer may also be filled */
928 	if (!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
929 	      (flip ? HCBUFSTAT_ISTL0_FULL : HCBUFSTAT_ISTL1_FULL))) {
930 		fno++;
931 		ptd_count = 0;
932 		flip = 1 - flip;
933 		goto fill2;
934 	}
935 }
936 
937 static void finish_transfers(struct isp1362_hcd *isp1362_hcd, unsigned long done_map,
938 			     struct isp1362_ep_queue *epq)
939 {
940 	struct isp1362_ep *ep;
941 	struct isp1362_ep *tmp;
942 
943 	if (list_empty(&epq->active)) {
944 		DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
945 		return;
946 	}
947 
948 	DBG(1, "%s: Finishing %s transfers %08lx\n", __func__, epq->name, done_map);
949 
950 	atomic_inc(&epq->finishing);
951 	list_for_each_entry_safe(ep, tmp, &epq->active, active) {
952 		int index = ep->ptd_index;
953 
954 		DBG(1, "%s: Checking %s PTD[%02x] $%04x\n", __func__, epq->name,
955 		    index, ep->ptd_offset);
956 
957 		BUG_ON(index < 0);
958 		if (__test_and_clear_bit(index, &done_map)) {
959 			isp1362_read_ptd(isp1362_hcd, ep, epq);
960 			epq->free_ptd = index;
961 			BUG_ON(ep->num_ptds == 0);
962 			release_ptd_buffers(epq, ep);
963 
964 			DBG(1, "%s: ep %p req %d removed from active list\n", __func__,
965 			    ep, ep->num_req);
966 			if (!list_empty(&ep->remove_list)) {
967 				list_del_init(&ep->remove_list);
968 				DBG(1, "%s: ep %p removed from remove list\n", __func__, ep);
969 			}
970 			DBG(1, "%s: Postprocessing %s ep %p req %d\n", __func__, epq->name,
971 			    ep, ep->num_req);
972 			postproc_ep(isp1362_hcd, ep);
973 		}
974 		if (!done_map)
975 			break;
976 	}
977 	if (done_map)
978 		pr_warn("%s: done_map not clear: %08lx:%08lx\n",
979 			__func__, done_map, epq->skip_map);
980 	atomic_dec(&epq->finishing);
981 }
982 
983 static void finish_iso_transfers(struct isp1362_hcd *isp1362_hcd, struct isp1362_ep_queue *epq)
984 {
985 	struct isp1362_ep *ep;
986 	struct isp1362_ep *tmp;
987 
988 	if (list_empty(&epq->active)) {
989 		DBG(1, "%s: Nothing to do for %s queue\n", __func__, epq->name);
990 		return;
991 	}
992 
993 	DBG(1, "%s: Finishing %s transfers\n", __func__, epq->name);
994 
995 	atomic_inc(&epq->finishing);
996 	list_for_each_entry_safe(ep, tmp, &epq->active, active) {
997 		DBG(1, "%s: Checking PTD $%04x\n", __func__, ep->ptd_offset);
998 
999 		isp1362_read_ptd(isp1362_hcd, ep, epq);
1000 		DBG(1, "%s: Postprocessing %s ep %p\n", __func__, epq->name, ep);
1001 		postproc_ep(isp1362_hcd, ep);
1002 	}
1003 	WARN_ON(epq->blk_size != 0);
1004 	atomic_dec(&epq->finishing);
1005 }
1006 
1007 static irqreturn_t isp1362_irq(struct usb_hcd *hcd)
1008 {
1009 	int handled = 0;
1010 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1011 	u16 irqstat;
1012 	u16 svc_mask;
1013 
1014 	spin_lock(&isp1362_hcd->lock);
1015 
1016 	BUG_ON(isp1362_hcd->irq_active++);
1017 
1018 	isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1019 
1020 	irqstat = isp1362_read_reg16(isp1362_hcd, HCuPINT);
1021 	DBG(3, "%s: got IRQ %04x:%04x\n", __func__, irqstat, isp1362_hcd->irqenb);
1022 
1023 	/* only handle interrupts that are currently enabled */
1024 	irqstat &= isp1362_hcd->irqenb;
1025 	isp1362_write_reg16(isp1362_hcd, HCuPINT, irqstat);
1026 	svc_mask = irqstat;
1027 
1028 	if (irqstat & HCuPINT_SOF) {
1029 		isp1362_hcd->irqenb &= ~HCuPINT_SOF;
1030 		isp1362_hcd->irq_stat[ISP1362_INT_SOF]++;
1031 		handled = 1;
1032 		svc_mask &= ~HCuPINT_SOF;
1033 		DBG(3, "%s: SOF\n", __func__);
1034 		isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1035 		if (!list_empty(&isp1362_hcd->remove_list))
1036 			finish_unlinks(isp1362_hcd);
1037 		if (!list_empty(&isp1362_hcd->async) && !(irqstat & HCuPINT_ATL)) {
1038 			if (list_empty(&isp1362_hcd->atl_queue.active)) {
1039 				start_atl_transfers(isp1362_hcd);
1040 			} else {
1041 				isp1362_enable_int(isp1362_hcd, HCuPINT_ATL);
1042 				isp1362_write_reg32(isp1362_hcd, HCATLSKIP,
1043 						    isp1362_hcd->atl_queue.skip_map);
1044 				isp1362_set_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1045 			}
1046 		}
1047 	}
1048 
1049 	if (irqstat & HCuPINT_ISTL0) {
1050 		isp1362_hcd->irq_stat[ISP1362_INT_ISTL0]++;
1051 		handled = 1;
1052 		svc_mask &= ~HCuPINT_ISTL0;
1053 		isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL0_FULL);
1054 		DBG(1, "%s: ISTL0\n", __func__);
1055 		WARN_ON((int)!!isp1362_hcd->istl_flip);
1056 		WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1057 			HCBUFSTAT_ISTL0_ACTIVE);
1058 		WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1059 			HCBUFSTAT_ISTL0_DONE));
1060 		isp1362_hcd->irqenb &= ~HCuPINT_ISTL0;
1061 	}
1062 
1063 	if (irqstat & HCuPINT_ISTL1) {
1064 		isp1362_hcd->irq_stat[ISP1362_INT_ISTL1]++;
1065 		handled = 1;
1066 		svc_mask &= ~HCuPINT_ISTL1;
1067 		isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ISTL1_FULL);
1068 		DBG(1, "%s: ISTL1\n", __func__);
1069 		WARN_ON(!(int)isp1362_hcd->istl_flip);
1070 		WARN_ON(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1071 			HCBUFSTAT_ISTL1_ACTIVE);
1072 		WARN_ON(!(isp1362_read_reg16(isp1362_hcd, HCBUFSTAT) &
1073 			HCBUFSTAT_ISTL1_DONE));
1074 		isp1362_hcd->irqenb &= ~HCuPINT_ISTL1;
1075 	}
1076 
1077 	if (irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) {
1078 		WARN_ON((irqstat & (HCuPINT_ISTL0 | HCuPINT_ISTL1)) ==
1079 			(HCuPINT_ISTL0 | HCuPINT_ISTL1));
1080 		finish_iso_transfers(isp1362_hcd,
1081 				     &isp1362_hcd->istl_queue[isp1362_hcd->istl_flip]);
1082 		start_iso_transfers(isp1362_hcd);
1083 		isp1362_hcd->istl_flip = 1 - isp1362_hcd->istl_flip;
1084 	}
1085 
1086 	if (irqstat & HCuPINT_INTL) {
1087 		u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1088 		u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCINTLSKIP);
1089 		isp1362_hcd->irq_stat[ISP1362_INT_INTL]++;
1090 
1091 		DBG(2, "%s: INTL\n", __func__);
1092 
1093 		svc_mask &= ~HCuPINT_INTL;
1094 
1095 		isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, skip_map | done_map);
1096 		if (~(done_map | skip_map) == 0)
1097 			/* All PTDs are finished, disable INTL processing entirely */
1098 			isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_INTL_ACTIVE);
1099 
1100 		handled = 1;
1101 		WARN_ON(!done_map);
1102 		if (done_map) {
1103 			DBG(3, "%s: INTL done_map %08x\n", __func__, done_map);
1104 			finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1105 			start_intl_transfers(isp1362_hcd);
1106 		}
1107 	}
1108 
1109 	if (irqstat & HCuPINT_ATL) {
1110 		u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1111 		u32 skip_map = isp1362_read_reg32(isp1362_hcd, HCATLSKIP);
1112 		isp1362_hcd->irq_stat[ISP1362_INT_ATL]++;
1113 
1114 		DBG(2, "%s: ATL\n", __func__);
1115 
1116 		svc_mask &= ~HCuPINT_ATL;
1117 
1118 		isp1362_write_reg32(isp1362_hcd, HCATLSKIP, skip_map | done_map);
1119 		if (~(done_map | skip_map) == 0)
1120 			isp1362_clr_mask16(isp1362_hcd, HCBUFSTAT, HCBUFSTAT_ATL_ACTIVE);
1121 		if (done_map) {
1122 			DBG(3, "%s: ATL done_map %08x\n", __func__, done_map);
1123 			finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1124 			start_atl_transfers(isp1362_hcd);
1125 		}
1126 		handled = 1;
1127 	}
1128 
1129 	if (irqstat & HCuPINT_OPR) {
1130 		u32 intstat = isp1362_read_reg32(isp1362_hcd, HCINTSTAT);
1131 		isp1362_hcd->irq_stat[ISP1362_INT_OPR]++;
1132 
1133 		svc_mask &= ~HCuPINT_OPR;
1134 		DBG(2, "%s: OPR %08x:%08x\n", __func__, intstat, isp1362_hcd->intenb);
1135 		intstat &= isp1362_hcd->intenb;
1136 		if (intstat & OHCI_INTR_UE) {
1137 			pr_err("Unrecoverable error\n");
1138 			/* FIXME: do here reset or cleanup or whatever */
1139 		}
1140 		if (intstat & OHCI_INTR_RHSC) {
1141 			isp1362_hcd->rhstatus = isp1362_read_reg32(isp1362_hcd, HCRHSTATUS);
1142 			isp1362_hcd->rhport[0] = isp1362_read_reg32(isp1362_hcd, HCRHPORT1);
1143 			isp1362_hcd->rhport[1] = isp1362_read_reg32(isp1362_hcd, HCRHPORT2);
1144 		}
1145 		if (intstat & OHCI_INTR_RD) {
1146 			pr_info("%s: RESUME DETECTED\n", __func__);
1147 			isp1362_show_reg(isp1362_hcd, HCCONTROL);
1148 			usb_hcd_resume_root_hub(hcd);
1149 		}
1150 		isp1362_write_reg32(isp1362_hcd, HCINTSTAT, intstat);
1151 		irqstat &= ~HCuPINT_OPR;
1152 		handled = 1;
1153 	}
1154 
1155 	if (irqstat & HCuPINT_SUSP) {
1156 		isp1362_hcd->irq_stat[ISP1362_INT_SUSP]++;
1157 		handled = 1;
1158 		svc_mask &= ~HCuPINT_SUSP;
1159 
1160 		pr_info("%s: SUSPEND IRQ\n", __func__);
1161 	}
1162 
1163 	if (irqstat & HCuPINT_CLKRDY) {
1164 		isp1362_hcd->irq_stat[ISP1362_INT_CLKRDY]++;
1165 		handled = 1;
1166 		isp1362_hcd->irqenb &= ~HCuPINT_CLKRDY;
1167 		svc_mask &= ~HCuPINT_CLKRDY;
1168 		pr_info("%s: CLKRDY IRQ\n", __func__);
1169 	}
1170 
1171 	if (svc_mask)
1172 		pr_err("%s: Unserviced interrupt(s) %04x\n", __func__, svc_mask);
1173 
1174 	isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
1175 	isp1362_hcd->irq_active--;
1176 	spin_unlock(&isp1362_hcd->lock);
1177 
1178 	return IRQ_RETVAL(handled);
1179 }
1180 
1181 /*-------------------------------------------------------------------------*/
1182 
1183 #define	MAX_PERIODIC_LOAD	900	/* out of 1000 usec */
1184 static int balance(struct isp1362_hcd *isp1362_hcd, u16 interval, u16 load)
1185 {
1186 	int i, branch = -ENOSPC;
1187 
1188 	/* search for the least loaded schedule branch of that interval
1189 	 * which has enough bandwidth left unreserved.
1190 	 */
1191 	for (i = 0; i < interval; i++) {
1192 		if (branch < 0 || isp1362_hcd->load[branch] > isp1362_hcd->load[i]) {
1193 			int j;
1194 
1195 			for (j = i; j < PERIODIC_SIZE; j += interval) {
1196 				if ((isp1362_hcd->load[j] + load) > MAX_PERIODIC_LOAD) {
1197 					pr_err("%s: new load %d load[%02x] %d max %d\n", __func__,
1198 					    load, j, isp1362_hcd->load[j], MAX_PERIODIC_LOAD);
1199 					break;
1200 				}
1201 			}
1202 			if (j < PERIODIC_SIZE)
1203 				continue;
1204 			branch = i;
1205 		}
1206 	}
1207 	return branch;
1208 }
1209 
1210 /* NB! ALL the code above this point runs with isp1362_hcd->lock
1211    held, irqs off
1212 */
1213 
1214 /*-------------------------------------------------------------------------*/
1215 
1216 static int isp1362_urb_enqueue(struct usb_hcd *hcd,
1217 			       struct urb *urb,
1218 			       gfp_t mem_flags)
1219 {
1220 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1221 	struct usb_device *udev = urb->dev;
1222 	unsigned int pipe = urb->pipe;
1223 	int is_out = !usb_pipein(pipe);
1224 	int type = usb_pipetype(pipe);
1225 	int epnum = usb_pipeendpoint(pipe);
1226 	struct usb_host_endpoint *hep = urb->ep;
1227 	struct isp1362_ep *ep = NULL;
1228 	unsigned long flags;
1229 	int retval = 0;
1230 
1231 	DBG(3, "%s: urb %p\n", __func__, urb);
1232 
1233 	if (type == PIPE_ISOCHRONOUS) {
1234 		pr_err("Isochronous transfers not supported\n");
1235 		return -ENOSPC;
1236 	}
1237 
1238 	URB_DBG("%s: FA %d ep%d%s %s: len %d %s%s\n", __func__,
1239 		usb_pipedevice(pipe), epnum,
1240 		is_out ? "out" : "in",
1241 		usb_pipecontrol(pipe) ? "ctrl" :
1242 			usb_pipeint(pipe) ? "int" :
1243 			usb_pipebulk(pipe) ? "bulk" :
1244 			"iso",
1245 		urb->transfer_buffer_length,
1246 		(urb->transfer_flags & URB_ZERO_PACKET) ? "ZERO_PACKET " : "",
1247 		!(urb->transfer_flags & URB_SHORT_NOT_OK) ?
1248 		"short_ok" : "");
1249 
1250 	/* avoid all allocations within spinlocks: request or endpoint */
1251 	if (!hep->hcpriv) {
1252 		ep = kzalloc(sizeof *ep, mem_flags);
1253 		if (!ep)
1254 			return -ENOMEM;
1255 	}
1256 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1257 
1258 	/* don't submit to a dead or disabled port */
1259 	if (!((isp1362_hcd->rhport[0] | isp1362_hcd->rhport[1]) &
1260 	      USB_PORT_STAT_ENABLE) ||
1261 	    !HC_IS_RUNNING(hcd->state)) {
1262 		kfree(ep);
1263 		retval = -ENODEV;
1264 		goto fail_not_linked;
1265 	}
1266 
1267 	retval = usb_hcd_link_urb_to_ep(hcd, urb);
1268 	if (retval) {
1269 		kfree(ep);
1270 		goto fail_not_linked;
1271 	}
1272 
1273 	if (hep->hcpriv) {
1274 		ep = hep->hcpriv;
1275 	} else {
1276 		INIT_LIST_HEAD(&ep->schedule);
1277 		INIT_LIST_HEAD(&ep->active);
1278 		INIT_LIST_HEAD(&ep->remove_list);
1279 		ep->udev = usb_get_dev(udev);
1280 		ep->hep = hep;
1281 		ep->epnum = epnum;
1282 		ep->maxpacket = usb_maxpacket(udev, urb->pipe);
1283 		ep->ptd_offset = -EINVAL;
1284 		ep->ptd_index = -EINVAL;
1285 		usb_settoggle(udev, epnum, is_out, 0);
1286 
1287 		if (type == PIPE_CONTROL)
1288 			ep->nextpid = USB_PID_SETUP;
1289 		else if (is_out)
1290 			ep->nextpid = USB_PID_OUT;
1291 		else
1292 			ep->nextpid = USB_PID_IN;
1293 
1294 		switch (type) {
1295 		case PIPE_ISOCHRONOUS:
1296 		case PIPE_INTERRUPT:
1297 			if (urb->interval > PERIODIC_SIZE)
1298 				urb->interval = PERIODIC_SIZE;
1299 			ep->interval = urb->interval;
1300 			ep->branch = PERIODIC_SIZE;
1301 			ep->load = usb_calc_bus_time(udev->speed, !is_out,
1302 						     type == PIPE_ISOCHRONOUS,
1303 						     usb_maxpacket(udev, pipe)) / 1000;
1304 			break;
1305 		}
1306 		hep->hcpriv = ep;
1307 	}
1308 	ep->num_req = isp1362_hcd->req_serial++;
1309 
1310 	/* maybe put endpoint into schedule */
1311 	switch (type) {
1312 	case PIPE_CONTROL:
1313 	case PIPE_BULK:
1314 		if (list_empty(&ep->schedule)) {
1315 			DBG(1, "%s: Adding ep %p req %d to async schedule\n",
1316 				__func__, ep, ep->num_req);
1317 			list_add_tail(&ep->schedule, &isp1362_hcd->async);
1318 		}
1319 		break;
1320 	case PIPE_ISOCHRONOUS:
1321 	case PIPE_INTERRUPT:
1322 		urb->interval = ep->interval;
1323 
1324 		/* urb submitted for already existing EP */
1325 		if (ep->branch < PERIODIC_SIZE)
1326 			break;
1327 
1328 		retval = balance(isp1362_hcd, ep->interval, ep->load);
1329 		if (retval < 0) {
1330 			pr_err("%s: balance returned %d\n", __func__, retval);
1331 			goto fail;
1332 		}
1333 		ep->branch = retval;
1334 		retval = 0;
1335 		isp1362_hcd->fmindex = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1336 		DBG(1, "%s: Current frame %04x branch %02x start_frame %04x(%04x)\n",
1337 		    __func__, isp1362_hcd->fmindex, ep->branch,
1338 		    ((isp1362_hcd->fmindex + PERIODIC_SIZE - 1) &
1339 		     ~(PERIODIC_SIZE - 1)) + ep->branch,
1340 		    (isp1362_hcd->fmindex & (PERIODIC_SIZE - 1)) + ep->branch);
1341 
1342 		if (list_empty(&ep->schedule)) {
1343 			if (type == PIPE_ISOCHRONOUS) {
1344 				u16 frame = isp1362_hcd->fmindex;
1345 
1346 				frame += max_t(u16, 8, ep->interval);
1347 				frame &= ~(ep->interval - 1);
1348 				frame |= ep->branch;
1349 				if (frame_before(frame, isp1362_hcd->fmindex))
1350 					frame += ep->interval;
1351 				urb->start_frame = frame;
1352 
1353 				DBG(1, "%s: Adding ep %p to isoc schedule\n", __func__, ep);
1354 				list_add_tail(&ep->schedule, &isp1362_hcd->isoc);
1355 			} else {
1356 				DBG(1, "%s: Adding ep %p to periodic schedule\n", __func__, ep);
1357 				list_add_tail(&ep->schedule, &isp1362_hcd->periodic);
1358 			}
1359 		} else
1360 			DBG(1, "%s: ep %p already scheduled\n", __func__, ep);
1361 
1362 		DBG(2, "%s: load %d bandwidth %d -> %d\n", __func__,
1363 		    ep->load / ep->interval, isp1362_hcd->load[ep->branch],
1364 		    isp1362_hcd->load[ep->branch] + ep->load);
1365 		isp1362_hcd->load[ep->branch] += ep->load;
1366 	}
1367 
1368 	urb->hcpriv = hep;
1369 	ALIGNSTAT(isp1362_hcd, urb->transfer_buffer);
1370 
1371 	switch (type) {
1372 	case PIPE_CONTROL:
1373 	case PIPE_BULK:
1374 		start_atl_transfers(isp1362_hcd);
1375 		break;
1376 	case PIPE_INTERRUPT:
1377 		start_intl_transfers(isp1362_hcd);
1378 		break;
1379 	case PIPE_ISOCHRONOUS:
1380 		start_iso_transfers(isp1362_hcd);
1381 		break;
1382 	default:
1383 		BUG();
1384 	}
1385  fail:
1386 	if (retval)
1387 		usb_hcd_unlink_urb_from_ep(hcd, urb);
1388 
1389 
1390  fail_not_linked:
1391 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1392 	if (retval)
1393 		DBG(0, "%s: urb %p failed with %d\n", __func__, urb, retval);
1394 	return retval;
1395 }
1396 
1397 static int isp1362_urb_dequeue(struct usb_hcd *hcd, struct urb *urb, int status)
1398 {
1399 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1400 	struct usb_host_endpoint *hep;
1401 	unsigned long flags;
1402 	struct isp1362_ep *ep;
1403 	int retval = 0;
1404 
1405 	DBG(3, "%s: urb %p\n", __func__, urb);
1406 
1407 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1408 	retval = usb_hcd_check_unlink_urb(hcd, urb, status);
1409 	if (retval)
1410 		goto done;
1411 
1412 	hep = urb->hcpriv;
1413 
1414 	if (!hep) {
1415 		spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1416 		return -EIDRM;
1417 	}
1418 
1419 	ep = hep->hcpriv;
1420 	if (ep) {
1421 		/* In front of queue? */
1422 		if (ep->hep->urb_list.next == &urb->urb_list) {
1423 			if (!list_empty(&ep->active)) {
1424 				DBG(1, "%s: urb %p ep %p req %d active PTD[%d] $%04x\n", __func__,
1425 				    urb, ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1426 				/* disable processing and queue PTD for removal */
1427 				remove_ptd(isp1362_hcd, ep);
1428 				urb = NULL;
1429 			}
1430 		}
1431 		if (urb) {
1432 			DBG(1, "%s: Finishing ep %p req %d\n", __func__, ep,
1433 			    ep->num_req);
1434 			finish_request(isp1362_hcd, ep, urb, status);
1435 		} else
1436 			DBG(1, "%s: urb %p active; wait4irq\n", __func__, urb);
1437 	} else {
1438 		pr_warn("%s: No EP in URB %p\n", __func__, urb);
1439 		retval = -EINVAL;
1440 	}
1441 done:
1442 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1443 
1444 	DBG(3, "%s: exit\n", __func__);
1445 
1446 	return retval;
1447 }
1448 
1449 static void isp1362_endpoint_disable(struct usb_hcd *hcd, struct usb_host_endpoint *hep)
1450 {
1451 	struct isp1362_ep *ep = hep->hcpriv;
1452 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1453 	unsigned long flags;
1454 
1455 	DBG(1, "%s: ep %p\n", __func__, ep);
1456 	if (!ep)
1457 		return;
1458 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1459 	if (!list_empty(&hep->urb_list)) {
1460 		if (!list_empty(&ep->active) && list_empty(&ep->remove_list)) {
1461 			DBG(1, "%s: Removing ep %p req %d PTD[%d] $%04x\n", __func__,
1462 			    ep, ep->num_req, ep->ptd_index, ep->ptd_offset);
1463 			remove_ptd(isp1362_hcd, ep);
1464 			pr_info("%s: Waiting for Interrupt to clean up\n", __func__);
1465 		}
1466 	}
1467 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1468 	/* Wait for interrupt to clear out active list */
1469 	while (!list_empty(&ep->active))
1470 		msleep(1);
1471 
1472 	DBG(1, "%s: Freeing EP %p\n", __func__, ep);
1473 
1474 	usb_put_dev(ep->udev);
1475 	kfree(ep);
1476 	hep->hcpriv = NULL;
1477 }
1478 
1479 static int isp1362_get_frame(struct usb_hcd *hcd)
1480 {
1481 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1482 	u32 fmnum;
1483 	unsigned long flags;
1484 
1485 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1486 	fmnum = isp1362_read_reg32(isp1362_hcd, HCFMNUM);
1487 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1488 
1489 	return (int)fmnum;
1490 }
1491 
1492 /*-------------------------------------------------------------------------*/
1493 
1494 /* Adapted from ohci-hub.c */
1495 static int isp1362_hub_status_data(struct usb_hcd *hcd, char *buf)
1496 {
1497 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1498 	int ports, i, changed = 0;
1499 	unsigned long flags;
1500 
1501 	if (!HC_IS_RUNNING(hcd->state))
1502 		return -ESHUTDOWN;
1503 
1504 	/* Report no status change now, if we are scheduled to be
1505 	   called later */
1506 	if (timer_pending(&hcd->rh_timer))
1507 		return 0;
1508 
1509 	ports = isp1362_hcd->rhdesca & RH_A_NDP;
1510 	BUG_ON(ports > 2);
1511 
1512 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1513 	/* init status */
1514 	if (isp1362_hcd->rhstatus & (RH_HS_LPSC | RH_HS_OCIC))
1515 		buf[0] = changed = 1;
1516 	else
1517 		buf[0] = 0;
1518 
1519 	for (i = 0; i < ports; i++) {
1520 		u32 status = isp1362_hcd->rhport[i];
1521 
1522 		if (status & (RH_PS_CSC | RH_PS_PESC | RH_PS_PSSC |
1523 			      RH_PS_OCIC | RH_PS_PRSC)) {
1524 			changed = 1;
1525 			buf[0] |= 1 << (i + 1);
1526 			continue;
1527 		}
1528 
1529 		if (!(status & RH_PS_CCS))
1530 			continue;
1531 	}
1532 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1533 	return changed;
1534 }
1535 
1536 static void isp1362_hub_descriptor(struct isp1362_hcd *isp1362_hcd,
1537 				   struct usb_hub_descriptor *desc)
1538 {
1539 	u32 reg = isp1362_hcd->rhdesca;
1540 
1541 	DBG(3, "%s: enter\n", __func__);
1542 
1543 	desc->bDescriptorType = USB_DT_HUB;
1544 	desc->bDescLength = 9;
1545 	desc->bHubContrCurrent = 0;
1546 	desc->bNbrPorts = reg & 0x3;
1547 	/* Power switching, device type, overcurrent. */
1548 	desc->wHubCharacteristics = cpu_to_le16((reg >> 8) &
1549 						(HUB_CHAR_LPSM |
1550 						 HUB_CHAR_COMPOUND |
1551 						 HUB_CHAR_OCPM));
1552 	DBG(0, "%s: hubcharacteristics = %02x\n", __func__,
1553 			desc->wHubCharacteristics);
1554 	desc->bPwrOn2PwrGood = (reg >> 24) & 0xff;
1555 	/* ports removable, and legacy PortPwrCtrlMask */
1556 	desc->u.hs.DeviceRemovable[0] = desc->bNbrPorts == 1 ? 1 << 1 : 3 << 1;
1557 	desc->u.hs.DeviceRemovable[1] = ~0;
1558 
1559 	DBG(3, "%s: exit\n", __func__);
1560 }
1561 
1562 /* Adapted from ohci-hub.c */
1563 static int isp1362_hub_control(struct usb_hcd *hcd, u16 typeReq, u16 wValue,
1564 			       u16 wIndex, char *buf, u16 wLength)
1565 {
1566 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1567 	int retval = 0;
1568 	unsigned long flags;
1569 	unsigned long t1;
1570 	int ports = isp1362_hcd->rhdesca & RH_A_NDP;
1571 	u32 tmp = 0;
1572 
1573 	switch (typeReq) {
1574 	case ClearHubFeature:
1575 		DBG(0, "ClearHubFeature: ");
1576 		switch (wValue) {
1577 		case C_HUB_OVER_CURRENT:
1578 			DBG(0, "C_HUB_OVER_CURRENT\n");
1579 			spin_lock_irqsave(&isp1362_hcd->lock, flags);
1580 			isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_OCIC);
1581 			spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1582 			break;
1583 		case C_HUB_LOCAL_POWER:
1584 			DBG(0, "C_HUB_LOCAL_POWER\n");
1585 			break;
1586 		default:
1587 			goto error;
1588 		}
1589 		break;
1590 	case SetHubFeature:
1591 		DBG(0, "SetHubFeature: ");
1592 		switch (wValue) {
1593 		case C_HUB_OVER_CURRENT:
1594 		case C_HUB_LOCAL_POWER:
1595 			DBG(0, "C_HUB_OVER_CURRENT or C_HUB_LOCAL_POWER\n");
1596 			break;
1597 		default:
1598 			goto error;
1599 		}
1600 		break;
1601 	case GetHubDescriptor:
1602 		DBG(0, "GetHubDescriptor\n");
1603 		isp1362_hub_descriptor(isp1362_hcd, (struct usb_hub_descriptor *)buf);
1604 		break;
1605 	case GetHubStatus:
1606 		DBG(0, "GetHubStatus\n");
1607 		put_unaligned(cpu_to_le32(0), (__le32 *) buf);
1608 		break;
1609 	case GetPortStatus:
1610 #ifndef VERBOSE
1611 		DBG(0, "GetPortStatus\n");
1612 #endif
1613 		if (!wIndex || wIndex > ports)
1614 			goto error;
1615 		tmp = isp1362_hcd->rhport[--wIndex];
1616 		put_unaligned(cpu_to_le32(tmp), (__le32 *) buf);
1617 		break;
1618 	case ClearPortFeature:
1619 		DBG(0, "ClearPortFeature: ");
1620 		if (!wIndex || wIndex > ports)
1621 			goto error;
1622 		wIndex--;
1623 
1624 		switch (wValue) {
1625 		case USB_PORT_FEAT_ENABLE:
1626 			DBG(0, "USB_PORT_FEAT_ENABLE\n");
1627 			tmp = RH_PS_CCS;
1628 			break;
1629 		case USB_PORT_FEAT_C_ENABLE:
1630 			DBG(0, "USB_PORT_FEAT_C_ENABLE\n");
1631 			tmp = RH_PS_PESC;
1632 			break;
1633 		case USB_PORT_FEAT_SUSPEND:
1634 			DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1635 			tmp = RH_PS_POCI;
1636 			break;
1637 		case USB_PORT_FEAT_C_SUSPEND:
1638 			DBG(0, "USB_PORT_FEAT_C_SUSPEND\n");
1639 			tmp = RH_PS_PSSC;
1640 			break;
1641 		case USB_PORT_FEAT_POWER:
1642 			DBG(0, "USB_PORT_FEAT_POWER\n");
1643 			tmp = RH_PS_LSDA;
1644 
1645 			break;
1646 		case USB_PORT_FEAT_C_CONNECTION:
1647 			DBG(0, "USB_PORT_FEAT_C_CONNECTION\n");
1648 			tmp = RH_PS_CSC;
1649 			break;
1650 		case USB_PORT_FEAT_C_OVER_CURRENT:
1651 			DBG(0, "USB_PORT_FEAT_C_OVER_CURRENT\n");
1652 			tmp = RH_PS_OCIC;
1653 			break;
1654 		case USB_PORT_FEAT_C_RESET:
1655 			DBG(0, "USB_PORT_FEAT_C_RESET\n");
1656 			tmp = RH_PS_PRSC;
1657 			break;
1658 		default:
1659 			goto error;
1660 		}
1661 
1662 		spin_lock_irqsave(&isp1362_hcd->lock, flags);
1663 		isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, tmp);
1664 		isp1362_hcd->rhport[wIndex] =
1665 			isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1666 		spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1667 		break;
1668 	case SetPortFeature:
1669 		DBG(0, "SetPortFeature: ");
1670 		if (!wIndex || wIndex > ports)
1671 			goto error;
1672 		wIndex--;
1673 		switch (wValue) {
1674 		case USB_PORT_FEAT_SUSPEND:
1675 			DBG(0, "USB_PORT_FEAT_SUSPEND\n");
1676 			spin_lock_irqsave(&isp1362_hcd->lock, flags);
1677 			isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PSS);
1678 			isp1362_hcd->rhport[wIndex] =
1679 				isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1680 			spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1681 			break;
1682 		case USB_PORT_FEAT_POWER:
1683 			DBG(0, "USB_PORT_FEAT_POWER\n");
1684 			spin_lock_irqsave(&isp1362_hcd->lock, flags);
1685 			isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, RH_PS_PPS);
1686 			isp1362_hcd->rhport[wIndex] =
1687 				isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1688 			spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1689 			break;
1690 		case USB_PORT_FEAT_RESET:
1691 			DBG(0, "USB_PORT_FEAT_RESET\n");
1692 			spin_lock_irqsave(&isp1362_hcd->lock, flags);
1693 
1694 			t1 = jiffies + msecs_to_jiffies(USB_RESET_WIDTH);
1695 			while (time_before(jiffies, t1)) {
1696 				/* spin until any current reset finishes */
1697 				for (;;) {
1698 					tmp = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + wIndex);
1699 					if (!(tmp & RH_PS_PRS))
1700 						break;
1701 					udelay(500);
1702 				}
1703 				if (!(tmp & RH_PS_CCS))
1704 					break;
1705 				/* Reset lasts 10ms (claims datasheet) */
1706 				isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + wIndex, (RH_PS_PRS));
1707 
1708 				spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1709 				msleep(10);
1710 				spin_lock_irqsave(&isp1362_hcd->lock, flags);
1711 			}
1712 
1713 			isp1362_hcd->rhport[wIndex] = isp1362_read_reg32(isp1362_hcd,
1714 									 HCRHPORT1 + wIndex);
1715 			spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1716 			break;
1717 		default:
1718 			goto error;
1719 		}
1720 		break;
1721 
1722 	default:
1723  error:
1724 		/* "protocol stall" on error */
1725 		DBG(0, "PROTOCOL STALL\n");
1726 		retval = -EPIPE;
1727 	}
1728 
1729 	return retval;
1730 }
1731 
1732 #ifdef	CONFIG_PM
1733 static int isp1362_bus_suspend(struct usb_hcd *hcd)
1734 {
1735 	int status = 0;
1736 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1737 	unsigned long flags;
1738 
1739 	if (time_before(jiffies, isp1362_hcd->next_statechange))
1740 		msleep(5);
1741 
1742 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1743 
1744 	isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1745 	switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1746 	case OHCI_USB_RESUME:
1747 		DBG(0, "%s: resume/suspend?\n", __func__);
1748 		isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1749 		isp1362_hcd->hc_control |= OHCI_USB_RESET;
1750 		isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1751 		fallthrough;
1752 	case OHCI_USB_RESET:
1753 		status = -EBUSY;
1754 		pr_warn("%s: needs reinit!\n", __func__);
1755 		goto done;
1756 	case OHCI_USB_SUSPEND:
1757 		pr_warn("%s: already suspended?\n", __func__);
1758 		goto done;
1759 	}
1760 	DBG(0, "%s: suspend root hub\n", __func__);
1761 
1762 	/* First stop any processing */
1763 	hcd->state = HC_STATE_QUIESCING;
1764 	if (!list_empty(&isp1362_hcd->atl_queue.active) ||
1765 	    !list_empty(&isp1362_hcd->intl_queue.active) ||
1766 	    !list_empty(&isp1362_hcd->istl_queue[0] .active) ||
1767 	    !list_empty(&isp1362_hcd->istl_queue[1] .active)) {
1768 		int limit;
1769 
1770 		isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
1771 		isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
1772 		isp1362_write_reg16(isp1362_hcd, HCBUFSTAT, 0);
1773 		isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
1774 		isp1362_write_reg32(isp1362_hcd, HCINTSTAT, OHCI_INTR_SF);
1775 
1776 		DBG(0, "%s: stopping schedules ...\n", __func__);
1777 		limit = 2000;
1778 		while (limit > 0) {
1779 			udelay(250);
1780 			limit -= 250;
1781 			if (isp1362_read_reg32(isp1362_hcd, HCINTSTAT) & OHCI_INTR_SF)
1782 				break;
1783 		}
1784 		mdelay(7);
1785 		if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ATL) {
1786 			u32 done_map = isp1362_read_reg32(isp1362_hcd, HCATLDONE);
1787 			finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->atl_queue);
1788 		}
1789 		if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_INTL) {
1790 			u32 done_map = isp1362_read_reg32(isp1362_hcd, HCINTLDONE);
1791 			finish_transfers(isp1362_hcd, done_map, &isp1362_hcd->intl_queue);
1792 		}
1793 		if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL0)
1794 			finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[0]);
1795 		if (isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_ISTL1)
1796 			finish_iso_transfers(isp1362_hcd, &isp1362_hcd->istl_queue[1]);
1797 	}
1798 	DBG(0, "%s: HCINTSTAT: %08x\n", __func__,
1799 		    isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1800 	isp1362_write_reg32(isp1362_hcd, HCINTSTAT,
1801 			    isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1802 
1803 	/* Suspend hub */
1804 	isp1362_hcd->hc_control = OHCI_USB_SUSPEND;
1805 	isp1362_show_reg(isp1362_hcd, HCCONTROL);
1806 	isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1807 	isp1362_show_reg(isp1362_hcd, HCCONTROL);
1808 
1809 #if 1
1810 	isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1811 	if ((isp1362_hcd->hc_control & OHCI_CTRL_HCFS) != OHCI_USB_SUSPEND) {
1812 		pr_err("%s: controller won't suspend %08x\n", __func__,
1813 		    isp1362_hcd->hc_control);
1814 		status = -EBUSY;
1815 	} else
1816 #endif
1817 	{
1818 		/* no resumes until devices finish suspending */
1819 		isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(5);
1820 	}
1821 done:
1822 	if (status == 0) {
1823 		hcd->state = HC_STATE_SUSPENDED;
1824 		DBG(0, "%s: HCD suspended: %08x\n", __func__,
1825 		    isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1826 	}
1827 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1828 	return status;
1829 }
1830 
1831 static int isp1362_bus_resume(struct usb_hcd *hcd)
1832 {
1833 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
1834 	u32 port;
1835 	unsigned long flags;
1836 	int status = -EINPROGRESS;
1837 
1838 	if (time_before(jiffies, isp1362_hcd->next_statechange))
1839 		msleep(5);
1840 
1841 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1842 	isp1362_hcd->hc_control = isp1362_read_reg32(isp1362_hcd, HCCONTROL);
1843 	pr_info("%s: HCCONTROL: %08x\n", __func__, isp1362_hcd->hc_control);
1844 	if (hcd->state == HC_STATE_RESUMING) {
1845 		pr_warn("%s: duplicate resume\n", __func__);
1846 		status = 0;
1847 	} else
1848 		switch (isp1362_hcd->hc_control & OHCI_CTRL_HCFS) {
1849 		case OHCI_USB_SUSPEND:
1850 			DBG(0, "%s: resume root hub\n", __func__);
1851 			isp1362_hcd->hc_control &= ~OHCI_CTRL_HCFS;
1852 			isp1362_hcd->hc_control |= OHCI_USB_RESUME;
1853 			isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1854 			break;
1855 		case OHCI_USB_RESUME:
1856 			/* HCFS changes sometime after INTR_RD */
1857 			DBG(0, "%s: remote wakeup\n", __func__);
1858 			break;
1859 		case OHCI_USB_OPER:
1860 			DBG(0, "%s: odd resume\n", __func__);
1861 			status = 0;
1862 			hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1863 			break;
1864 		default:		/* RESET, we lost power */
1865 			DBG(0, "%s: root hub hardware reset\n", __func__);
1866 			status = -EBUSY;
1867 		}
1868 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1869 	if (status == -EBUSY) {
1870 		DBG(0, "%s: Restarting HC\n", __func__);
1871 		isp1362_hc_stop(hcd);
1872 		return isp1362_hc_start(hcd);
1873 	}
1874 	if (status != -EINPROGRESS)
1875 		return status;
1876 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1877 	port = isp1362_read_reg32(isp1362_hcd, HCRHDESCA) & RH_A_NDP;
1878 	while (port--) {
1879 		u32 stat = isp1362_read_reg32(isp1362_hcd, HCRHPORT1 + port);
1880 
1881 		/* force global, not selective, resume */
1882 		if (!(stat & RH_PS_PSS)) {
1883 			DBG(0, "%s: Not Resuming RH port %d\n", __func__, port);
1884 			continue;
1885 		}
1886 		DBG(0, "%s: Resuming RH port %d\n", __func__, port);
1887 		isp1362_write_reg32(isp1362_hcd, HCRHPORT1 + port, RH_PS_POCI);
1888 	}
1889 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1890 
1891 	/* Some controllers (lucent) need extra-long delays */
1892 	hcd->state = HC_STATE_RESUMING;
1893 	mdelay(20 /* usb 11.5.1.10 */ + 15);
1894 
1895 	isp1362_hcd->hc_control = OHCI_USB_OPER;
1896 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
1897 	isp1362_show_reg(isp1362_hcd, HCCONTROL);
1898 	isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
1899 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
1900 	/* TRSMRCY */
1901 	msleep(10);
1902 
1903 	/* keep it alive for ~5x suspend + resume costs */
1904 	isp1362_hcd->next_statechange = jiffies + msecs_to_jiffies(250);
1905 
1906 	hcd->self.root_hub->dev.power.power_state = PMSG_ON;
1907 	hcd->state = HC_STATE_RUNNING;
1908 	return 0;
1909 }
1910 #else
1911 #define	isp1362_bus_suspend	NULL
1912 #define	isp1362_bus_resume	NULL
1913 #endif
1914 
1915 /*-------------------------------------------------------------------------*/
1916 
1917 static void dump_irq(struct seq_file *s, char *label, u16 mask)
1918 {
1919 	seq_printf(s, "%-15s %04x%s%s%s%s%s%s\n", label, mask,
1920 		   mask & HCuPINT_CLKRDY ? " clkrdy" : "",
1921 		   mask & HCuPINT_SUSP ? " susp" : "",
1922 		   mask & HCuPINT_OPR ? " opr" : "",
1923 		   mask & HCuPINT_EOT ? " eot" : "",
1924 		   mask & HCuPINT_ATL ? " atl" : "",
1925 		   mask & HCuPINT_SOF ? " sof" : "");
1926 }
1927 
1928 static void dump_int(struct seq_file *s, char *label, u32 mask)
1929 {
1930 	seq_printf(s, "%-15s %08x%s%s%s%s%s%s%s\n", label, mask,
1931 		   mask & OHCI_INTR_MIE ? " MIE" : "",
1932 		   mask & OHCI_INTR_RHSC ? " rhsc" : "",
1933 		   mask & OHCI_INTR_FNO ? " fno" : "",
1934 		   mask & OHCI_INTR_UE ? " ue" : "",
1935 		   mask & OHCI_INTR_RD ? " rd" : "",
1936 		   mask & OHCI_INTR_SF ? " sof" : "",
1937 		   mask & OHCI_INTR_SO ? " so" : "");
1938 }
1939 
1940 static void dump_ctrl(struct seq_file *s, char *label, u32 mask)
1941 {
1942 	seq_printf(s, "%-15s %08x%s%s%s\n", label, mask,
1943 		   mask & OHCI_CTRL_RWC ? " rwc" : "",
1944 		   mask & OHCI_CTRL_RWE ? " rwe" : "",
1945 		   ({
1946 			   char *hcfs;
1947 			   switch (mask & OHCI_CTRL_HCFS) {
1948 			   case OHCI_USB_OPER:
1949 				   hcfs = " oper";
1950 				   break;
1951 			   case OHCI_USB_RESET:
1952 				   hcfs = " reset";
1953 				   break;
1954 			   case OHCI_USB_RESUME:
1955 				   hcfs = " resume";
1956 				   break;
1957 			   case OHCI_USB_SUSPEND:
1958 				   hcfs = " suspend";
1959 				   break;
1960 			   default:
1961 				   hcfs = " ?";
1962 			   }
1963 			   hcfs;
1964 		   }));
1965 }
1966 
1967 static void dump_regs(struct seq_file *s, struct isp1362_hcd *isp1362_hcd)
1968 {
1969 	seq_printf(s, "HCREVISION [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCREVISION),
1970 		   isp1362_read_reg32(isp1362_hcd, HCREVISION));
1971 	seq_printf(s, "HCCONTROL  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCONTROL),
1972 		   isp1362_read_reg32(isp1362_hcd, HCCONTROL));
1973 	seq_printf(s, "HCCMDSTAT  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCCMDSTAT),
1974 		   isp1362_read_reg32(isp1362_hcd, HCCMDSTAT));
1975 	seq_printf(s, "HCINTSTAT  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTSTAT),
1976 		   isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
1977 	seq_printf(s, "HCINTENB   [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTENB),
1978 		   isp1362_read_reg32(isp1362_hcd, HCINTENB));
1979 	seq_printf(s, "HCFMINTVL  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMINTVL),
1980 		   isp1362_read_reg32(isp1362_hcd, HCFMINTVL));
1981 	seq_printf(s, "HCFMREM    [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMREM),
1982 		   isp1362_read_reg32(isp1362_hcd, HCFMREM));
1983 	seq_printf(s, "HCFMNUM    [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCFMNUM),
1984 		   isp1362_read_reg32(isp1362_hcd, HCFMNUM));
1985 	seq_printf(s, "HCLSTHRESH [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCLSTHRESH),
1986 		   isp1362_read_reg32(isp1362_hcd, HCLSTHRESH));
1987 	seq_printf(s, "HCRHDESCA  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCA),
1988 		   isp1362_read_reg32(isp1362_hcd, HCRHDESCA));
1989 	seq_printf(s, "HCRHDESCB  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHDESCB),
1990 		   isp1362_read_reg32(isp1362_hcd, HCRHDESCB));
1991 	seq_printf(s, "HCRHSTATUS [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHSTATUS),
1992 		   isp1362_read_reg32(isp1362_hcd, HCRHSTATUS));
1993 	seq_printf(s, "HCRHPORT1  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT1),
1994 		   isp1362_read_reg32(isp1362_hcd, HCRHPORT1));
1995 	seq_printf(s, "HCRHPORT2  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCRHPORT2),
1996 		   isp1362_read_reg32(isp1362_hcd, HCRHPORT2));
1997 	seq_printf(s, "\n");
1998 	seq_printf(s, "HCHWCFG    [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCHWCFG),
1999 		   isp1362_read_reg16(isp1362_hcd, HCHWCFG));
2000 	seq_printf(s, "HCDMACFG   [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCDMACFG),
2001 		   isp1362_read_reg16(isp1362_hcd, HCDMACFG));
2002 	seq_printf(s, "HCXFERCTR  [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCXFERCTR),
2003 		   isp1362_read_reg16(isp1362_hcd, HCXFERCTR));
2004 	seq_printf(s, "HCuPINT    [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINT),
2005 		   isp1362_read_reg16(isp1362_hcd, HCuPINT));
2006 	seq_printf(s, "HCuPINTENB [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCuPINTENB),
2007 		   isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2008 	seq_printf(s, "HCCHIPID   [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCCHIPID),
2009 		   isp1362_read_reg16(isp1362_hcd, HCCHIPID));
2010 	seq_printf(s, "HCSCRATCH  [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCSCRATCH),
2011 		   isp1362_read_reg16(isp1362_hcd, HCSCRATCH));
2012 	seq_printf(s, "HCBUFSTAT  [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCBUFSTAT),
2013 		   isp1362_read_reg16(isp1362_hcd, HCBUFSTAT));
2014 	seq_printf(s, "HCDIRADDR  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCDIRADDR),
2015 		   isp1362_read_reg32(isp1362_hcd, HCDIRADDR));
2016 #if 0
2017 	seq_printf(s, "HCDIRDATA  [%02x]     %04x\n", ISP1362_REG_NO(HCDIRDATA),
2018 		   isp1362_read_reg16(isp1362_hcd, HCDIRDATA));
2019 #endif
2020 	seq_printf(s, "HCISTLBUFSZ[%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLBUFSZ),
2021 		   isp1362_read_reg16(isp1362_hcd, HCISTLBUFSZ));
2022 	seq_printf(s, "HCISTLRATE [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCISTLRATE),
2023 		   isp1362_read_reg16(isp1362_hcd, HCISTLRATE));
2024 	seq_printf(s, "\n");
2025 	seq_printf(s, "HCINTLBUFSZ[%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBUFSZ),
2026 		   isp1362_read_reg16(isp1362_hcd, HCINTLBUFSZ));
2027 	seq_printf(s, "HCINTLBLKSZ[%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLBLKSZ),
2028 		   isp1362_read_reg16(isp1362_hcd, HCINTLBLKSZ));
2029 	seq_printf(s, "HCINTLDONE [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLDONE),
2030 		   isp1362_read_reg32(isp1362_hcd, HCINTLDONE));
2031 	seq_printf(s, "HCINTLSKIP [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLSKIP),
2032 		   isp1362_read_reg32(isp1362_hcd, HCINTLSKIP));
2033 	seq_printf(s, "HCINTLLAST [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLLAST),
2034 		   isp1362_read_reg32(isp1362_hcd, HCINTLLAST));
2035 	seq_printf(s, "HCINTLCURR [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCINTLCURR),
2036 		   isp1362_read_reg16(isp1362_hcd, HCINTLCURR));
2037 	seq_printf(s, "\n");
2038 	seq_printf(s, "HCATLBUFSZ [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBUFSZ),
2039 		   isp1362_read_reg16(isp1362_hcd, HCATLBUFSZ));
2040 	seq_printf(s, "HCATLBLKSZ [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLBLKSZ),
2041 		   isp1362_read_reg16(isp1362_hcd, HCATLBLKSZ));
2042 #if 0
2043 	seq_printf(s, "HCATLDONE  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDONE),
2044 		   isp1362_read_reg32(isp1362_hcd, HCATLDONE));
2045 #endif
2046 	seq_printf(s, "HCATLSKIP  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLSKIP),
2047 		   isp1362_read_reg32(isp1362_hcd, HCATLSKIP));
2048 	seq_printf(s, "HCATLLAST  [%02x] %08x\n", ISP1362_REG_NO(ISP1362_REG_HCATLLAST),
2049 		   isp1362_read_reg32(isp1362_hcd, HCATLLAST));
2050 	seq_printf(s, "HCATLCURR  [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLCURR),
2051 		   isp1362_read_reg16(isp1362_hcd, HCATLCURR));
2052 	seq_printf(s, "\n");
2053 	seq_printf(s, "HCATLDTC   [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTC),
2054 		   isp1362_read_reg16(isp1362_hcd, HCATLDTC));
2055 	seq_printf(s, "HCATLDTCTO [%02x]     %04x\n", ISP1362_REG_NO(ISP1362_REG_HCATLDTCTO),
2056 		   isp1362_read_reg16(isp1362_hcd, HCATLDTCTO));
2057 }
2058 
2059 static int isp1362_show(struct seq_file *s, void *unused)
2060 {
2061 	struct isp1362_hcd *isp1362_hcd = s->private;
2062 	struct isp1362_ep *ep;
2063 	int i;
2064 
2065 	seq_printf(s, "%s\n%s version %s\n",
2066 		   isp1362_hcd_to_hcd(isp1362_hcd)->product_desc, hcd_name, DRIVER_VERSION);
2067 
2068 	/* collect statistics to help estimate potential win for
2069 	 * DMA engines that care about alignment (PXA)
2070 	 */
2071 	seq_printf(s, "alignment:  16b/%ld 8b/%ld 4b/%ld 2b/%ld 1b/%ld\n",
2072 		   isp1362_hcd->stat16, isp1362_hcd->stat8, isp1362_hcd->stat4,
2073 		   isp1362_hcd->stat2, isp1362_hcd->stat1);
2074 	seq_printf(s, "max # ptds in ATL  fifo: %d\n", isp1362_hcd->atl_queue.stat_maxptds);
2075 	seq_printf(s, "max # ptds in INTL fifo: %d\n", isp1362_hcd->intl_queue.stat_maxptds);
2076 	seq_printf(s, "max # ptds in ISTL fifo: %d\n",
2077 		   max(isp1362_hcd->istl_queue[0] .stat_maxptds,
2078 		       isp1362_hcd->istl_queue[1] .stat_maxptds));
2079 
2080 	/* FIXME: don't show the following in suspended state */
2081 	spin_lock_irq(&isp1362_hcd->lock);
2082 
2083 	dump_irq(s, "hc_irq_enable", isp1362_read_reg16(isp1362_hcd, HCuPINTENB));
2084 	dump_irq(s, "hc_irq_status", isp1362_read_reg16(isp1362_hcd, HCuPINT));
2085 	dump_int(s, "ohci_int_enable", isp1362_read_reg32(isp1362_hcd, HCINTENB));
2086 	dump_int(s, "ohci_int_status", isp1362_read_reg32(isp1362_hcd, HCINTSTAT));
2087 	dump_ctrl(s, "ohci_control", isp1362_read_reg32(isp1362_hcd, HCCONTROL));
2088 
2089 	for (i = 0; i < NUM_ISP1362_IRQS; i++)
2090 		if (isp1362_hcd->irq_stat[i])
2091 			seq_printf(s, "%-15s: %d\n",
2092 				   ISP1362_INT_NAME(i), isp1362_hcd->irq_stat[i]);
2093 
2094 	dump_regs(s, isp1362_hcd);
2095 	list_for_each_entry(ep, &isp1362_hcd->async, schedule) {
2096 		struct urb *urb;
2097 
2098 		seq_printf(s, "%p, ep%d%s, maxpacket %d:\n", ep, ep->epnum,
2099 			   ({
2100 				   char *s;
2101 				   switch (ep->nextpid) {
2102 				   case USB_PID_IN:
2103 					   s = "in";
2104 					   break;
2105 				   case USB_PID_OUT:
2106 					   s = "out";
2107 					   break;
2108 				   case USB_PID_SETUP:
2109 					   s = "setup";
2110 					   break;
2111 				   case USB_PID_ACK:
2112 					   s = "status";
2113 					   break;
2114 				   default:
2115 					   s = "?";
2116 					   break;
2117 				   }
2118 				   s;}), ep->maxpacket) ;
2119 		list_for_each_entry(urb, &ep->hep->urb_list, urb_list) {
2120 			seq_printf(s, "  urb%p, %d/%d\n", urb,
2121 				   urb->actual_length,
2122 				   urb->transfer_buffer_length);
2123 		}
2124 	}
2125 	if (!list_empty(&isp1362_hcd->async))
2126 		seq_printf(s, "\n");
2127 	dump_ptd_queue(&isp1362_hcd->atl_queue);
2128 
2129 	seq_printf(s, "periodic size= %d\n", PERIODIC_SIZE);
2130 
2131 	list_for_each_entry(ep, &isp1362_hcd->periodic, schedule) {
2132 		seq_printf(s, "branch:%2d load:%3d PTD[%d] $%04x:\n", ep->branch,
2133 			   isp1362_hcd->load[ep->branch], ep->ptd_index, ep->ptd_offset);
2134 
2135 		seq_printf(s, "   %d/%p (%sdev%d ep%d%s max %d)\n",
2136 			   ep->interval, ep,
2137 			   (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2138 			   ep->udev->devnum, ep->epnum,
2139 			   (ep->epnum == 0) ? "" :
2140 			   ((ep->nextpid == USB_PID_IN) ?
2141 			    "in" : "out"), ep->maxpacket);
2142 	}
2143 	dump_ptd_queue(&isp1362_hcd->intl_queue);
2144 
2145 	seq_printf(s, "ISO:\n");
2146 
2147 	list_for_each_entry(ep, &isp1362_hcd->isoc, schedule) {
2148 		seq_printf(s, "   %d/%p (%sdev%d ep%d%s max %d)\n",
2149 			   ep->interval, ep,
2150 			   (ep->udev->speed == USB_SPEED_FULL) ? "" : "ls ",
2151 			   ep->udev->devnum, ep->epnum,
2152 			   (ep->epnum == 0) ? "" :
2153 			   ((ep->nextpid == USB_PID_IN) ?
2154 			    "in" : "out"), ep->maxpacket);
2155 	}
2156 
2157 	spin_unlock_irq(&isp1362_hcd->lock);
2158 	seq_printf(s, "\n");
2159 
2160 	return 0;
2161 }
2162 DEFINE_SHOW_ATTRIBUTE(isp1362);
2163 
2164 /* expect just one isp1362_hcd per system */
2165 static void create_debug_file(struct isp1362_hcd *isp1362_hcd)
2166 {
2167 	debugfs_create_file("isp1362", S_IRUGO, usb_debug_root, isp1362_hcd,
2168 			    &isp1362_fops);
2169 }
2170 
2171 static void remove_debug_file(struct isp1362_hcd *isp1362_hcd)
2172 {
2173 	debugfs_remove(debugfs_lookup("isp1362", usb_debug_root));
2174 }
2175 
2176 /*-------------------------------------------------------------------------*/
2177 
2178 static void __isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2179 {
2180 	int tmp = 20;
2181 
2182 	isp1362_write_reg16(isp1362_hcd, HCSWRES, HCSWRES_MAGIC);
2183 	isp1362_write_reg32(isp1362_hcd, HCCMDSTAT, OHCI_HCR);
2184 	while (--tmp) {
2185 		mdelay(1);
2186 		if (!(isp1362_read_reg32(isp1362_hcd, HCCMDSTAT) & OHCI_HCR))
2187 			break;
2188 	}
2189 	if (!tmp)
2190 		pr_err("Software reset timeout\n");
2191 }
2192 
2193 static void isp1362_sw_reset(struct isp1362_hcd *isp1362_hcd)
2194 {
2195 	unsigned long flags;
2196 
2197 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2198 	__isp1362_sw_reset(isp1362_hcd);
2199 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2200 }
2201 
2202 static int isp1362_mem_config(struct usb_hcd *hcd)
2203 {
2204 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2205 	unsigned long flags;
2206 	u32 total;
2207 	u16 istl_size = ISP1362_ISTL_BUFSIZE;
2208 	u16 intl_blksize = ISP1362_INTL_BLKSIZE + PTD_HEADER_SIZE;
2209 	u16 intl_size = ISP1362_INTL_BUFFERS * intl_blksize;
2210 	u16 atl_blksize = ISP1362_ATL_BLKSIZE + PTD_HEADER_SIZE;
2211 	u16 atl_buffers = (ISP1362_BUF_SIZE - (istl_size + intl_size)) / atl_blksize;
2212 	u16 atl_size;
2213 	int i;
2214 
2215 	WARN_ON(istl_size & 3);
2216 	WARN_ON(atl_blksize & 3);
2217 	WARN_ON(intl_blksize & 3);
2218 	WARN_ON(atl_blksize < PTD_HEADER_SIZE);
2219 	WARN_ON(intl_blksize < PTD_HEADER_SIZE);
2220 
2221 	BUG_ON((unsigned)ISP1362_INTL_BUFFERS > 32);
2222 	if (atl_buffers > 32)
2223 		atl_buffers = 32;
2224 	atl_size = atl_buffers * atl_blksize;
2225 	total = atl_size + intl_size + istl_size;
2226 	dev_info(hcd->self.controller, "ISP1362 Memory usage:\n");
2227 	dev_info(hcd->self.controller, "  ISTL:    2 * %4d:     %4d @ $%04x:$%04x\n",
2228 		 istl_size / 2, istl_size, 0, istl_size / 2);
2229 	dev_info(hcd->self.controller, "  INTL: %4d * (%3zu+8):  %4d @ $%04x\n",
2230 		 ISP1362_INTL_BUFFERS, intl_blksize - PTD_HEADER_SIZE,
2231 		 intl_size, istl_size);
2232 	dev_info(hcd->self.controller, "  ATL : %4d * (%3zu+8):  %4d @ $%04x\n",
2233 		 atl_buffers, atl_blksize - PTD_HEADER_SIZE,
2234 		 atl_size, istl_size + intl_size);
2235 	dev_info(hcd->self.controller, "  USED/FREE:   %4d      %4d\n", total,
2236 		 ISP1362_BUF_SIZE - total);
2237 
2238 	if (total > ISP1362_BUF_SIZE) {
2239 		dev_err(hcd->self.controller, "%s: Memory requested: %d, available %d\n",
2240 			__func__, total, ISP1362_BUF_SIZE);
2241 		return -ENOMEM;
2242 	}
2243 
2244 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2245 
2246 	for (i = 0; i < 2; i++) {
2247 		isp1362_hcd->istl_queue[i].buf_start = i * istl_size / 2,
2248 		isp1362_hcd->istl_queue[i].buf_size = istl_size / 2;
2249 		isp1362_hcd->istl_queue[i].blk_size = 4;
2250 		INIT_LIST_HEAD(&isp1362_hcd->istl_queue[i].active);
2251 		snprintf(isp1362_hcd->istl_queue[i].name,
2252 			 sizeof(isp1362_hcd->istl_queue[i].name), "ISTL%d", i);
2253 		DBG(3, "%s: %5s buf $%04x %d\n", __func__,
2254 		     isp1362_hcd->istl_queue[i].name,
2255 		     isp1362_hcd->istl_queue[i].buf_start,
2256 		     isp1362_hcd->istl_queue[i].buf_size);
2257 	}
2258 	isp1362_write_reg16(isp1362_hcd, HCISTLBUFSZ, istl_size / 2);
2259 
2260 	isp1362_hcd->intl_queue.buf_start = istl_size;
2261 	isp1362_hcd->intl_queue.buf_size = intl_size;
2262 	isp1362_hcd->intl_queue.buf_count = ISP1362_INTL_BUFFERS;
2263 	isp1362_hcd->intl_queue.blk_size = intl_blksize;
2264 	isp1362_hcd->intl_queue.buf_avail = isp1362_hcd->intl_queue.buf_count;
2265 	isp1362_hcd->intl_queue.skip_map = ~0;
2266 	INIT_LIST_HEAD(&isp1362_hcd->intl_queue.active);
2267 
2268 	isp1362_write_reg16(isp1362_hcd, HCINTLBUFSZ,
2269 			    isp1362_hcd->intl_queue.buf_size);
2270 	isp1362_write_reg16(isp1362_hcd, HCINTLBLKSZ,
2271 			    isp1362_hcd->intl_queue.blk_size - PTD_HEADER_SIZE);
2272 	isp1362_write_reg32(isp1362_hcd, HCINTLSKIP, ~0);
2273 	isp1362_write_reg32(isp1362_hcd, HCINTLLAST,
2274 			    1 << (ISP1362_INTL_BUFFERS - 1));
2275 
2276 	isp1362_hcd->atl_queue.buf_start = istl_size + intl_size;
2277 	isp1362_hcd->atl_queue.buf_size = atl_size;
2278 	isp1362_hcd->atl_queue.buf_count = atl_buffers;
2279 	isp1362_hcd->atl_queue.blk_size = atl_blksize;
2280 	isp1362_hcd->atl_queue.buf_avail = isp1362_hcd->atl_queue.buf_count;
2281 	isp1362_hcd->atl_queue.skip_map = ~0;
2282 	INIT_LIST_HEAD(&isp1362_hcd->atl_queue.active);
2283 
2284 	isp1362_write_reg16(isp1362_hcd, HCATLBUFSZ,
2285 			    isp1362_hcd->atl_queue.buf_size);
2286 	isp1362_write_reg16(isp1362_hcd, HCATLBLKSZ,
2287 			    isp1362_hcd->atl_queue.blk_size - PTD_HEADER_SIZE);
2288 	isp1362_write_reg32(isp1362_hcd, HCATLSKIP, ~0);
2289 	isp1362_write_reg32(isp1362_hcd, HCATLLAST,
2290 			    1 << (atl_buffers - 1));
2291 
2292 	snprintf(isp1362_hcd->atl_queue.name,
2293 		 sizeof(isp1362_hcd->atl_queue.name), "ATL");
2294 	snprintf(isp1362_hcd->intl_queue.name,
2295 		 sizeof(isp1362_hcd->intl_queue.name), "INTL");
2296 	DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2297 	     isp1362_hcd->intl_queue.name,
2298 	     isp1362_hcd->intl_queue.buf_start,
2299 	     ISP1362_INTL_BUFFERS, isp1362_hcd->intl_queue.blk_size,
2300 	     isp1362_hcd->intl_queue.buf_size);
2301 	DBG(3, "%s: %5s buf $%04x %2d * %4d = %4d\n", __func__,
2302 	     isp1362_hcd->atl_queue.name,
2303 	     isp1362_hcd->atl_queue.buf_start,
2304 	     atl_buffers, isp1362_hcd->atl_queue.blk_size,
2305 	     isp1362_hcd->atl_queue.buf_size);
2306 
2307 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2308 
2309 	return 0;
2310 }
2311 
2312 static int isp1362_hc_reset(struct usb_hcd *hcd)
2313 {
2314 	int ret = 0;
2315 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2316 	unsigned long t;
2317 	unsigned long timeout = 100;
2318 	unsigned long flags;
2319 	int clkrdy = 0;
2320 
2321 	pr_debug("%s:\n", __func__);
2322 
2323 	if (isp1362_hcd->board && isp1362_hcd->board->reset) {
2324 		isp1362_hcd->board->reset(hcd->self.controller, 1);
2325 		msleep(20);
2326 		if (isp1362_hcd->board->clock)
2327 			isp1362_hcd->board->clock(hcd->self.controller, 1);
2328 		isp1362_hcd->board->reset(hcd->self.controller, 0);
2329 	} else
2330 		isp1362_sw_reset(isp1362_hcd);
2331 
2332 	/* chip has been reset. First we need to see a clock */
2333 	t = jiffies + msecs_to_jiffies(timeout);
2334 	while (!clkrdy && time_before_eq(jiffies, t)) {
2335 		spin_lock_irqsave(&isp1362_hcd->lock, flags);
2336 		clkrdy = isp1362_read_reg16(isp1362_hcd, HCuPINT) & HCuPINT_CLKRDY;
2337 		spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2338 		if (!clkrdy)
2339 			msleep(4);
2340 	}
2341 
2342 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2343 	isp1362_write_reg16(isp1362_hcd, HCuPINT, HCuPINT_CLKRDY);
2344 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2345 	if (!clkrdy) {
2346 		pr_err("Clock not ready after %lums\n", timeout);
2347 		ret = -ENODEV;
2348 	}
2349 	return ret;
2350 }
2351 
2352 static void isp1362_hc_stop(struct usb_hcd *hcd)
2353 {
2354 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2355 	unsigned long flags;
2356 	u32 tmp;
2357 
2358 	pr_debug("%s:\n", __func__);
2359 
2360 	del_timer_sync(&hcd->rh_timer);
2361 
2362 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2363 
2364 	isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2365 
2366 	/* Switch off power for all ports */
2367 	tmp = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2368 	tmp &= ~(RH_A_NPS | RH_A_PSM);
2369 	isp1362_write_reg32(isp1362_hcd, HCRHDESCA, tmp);
2370 	isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2371 
2372 	/* Reset the chip */
2373 	if (isp1362_hcd->board && isp1362_hcd->board->reset)
2374 		isp1362_hcd->board->reset(hcd->self.controller, 1);
2375 	else
2376 		__isp1362_sw_reset(isp1362_hcd);
2377 
2378 	if (isp1362_hcd->board && isp1362_hcd->board->clock)
2379 		isp1362_hcd->board->clock(hcd->self.controller, 0);
2380 
2381 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2382 }
2383 
2384 #ifdef CHIP_BUFFER_TEST
2385 static int isp1362_chip_test(struct isp1362_hcd *isp1362_hcd)
2386 {
2387 	int ret = 0;
2388 	u16 *ref;
2389 	unsigned long flags;
2390 
2391 	ref = kmalloc(2 * ISP1362_BUF_SIZE, GFP_KERNEL);
2392 	if (ref) {
2393 		int offset;
2394 		u16 *tst = &ref[ISP1362_BUF_SIZE / 2];
2395 
2396 		for (offset = 0; offset < ISP1362_BUF_SIZE / 2; offset++) {
2397 			ref[offset] = ~offset;
2398 			tst[offset] = offset;
2399 		}
2400 
2401 		for (offset = 0; offset < 4; offset++) {
2402 			int j;
2403 
2404 			for (j = 0; j < 8; j++) {
2405 				spin_lock_irqsave(&isp1362_hcd->lock, flags);
2406 				isp1362_write_buffer(isp1362_hcd, (u8 *)ref + offset, 0, j);
2407 				isp1362_read_buffer(isp1362_hcd, (u8 *)tst + offset, 0, j);
2408 				spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2409 
2410 				if (memcmp(ref, tst, j)) {
2411 					ret = -ENODEV;
2412 					pr_err("%s: memory check with %d byte offset %d failed\n",
2413 					    __func__, j, offset);
2414 					dump_data((u8 *)ref + offset, j);
2415 					dump_data((u8 *)tst + offset, j);
2416 				}
2417 			}
2418 		}
2419 
2420 		spin_lock_irqsave(&isp1362_hcd->lock, flags);
2421 		isp1362_write_buffer(isp1362_hcd, ref, 0, ISP1362_BUF_SIZE);
2422 		isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2423 		spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2424 
2425 		if (memcmp(ref, tst, ISP1362_BUF_SIZE)) {
2426 			ret = -ENODEV;
2427 			pr_err("%s: memory check failed\n", __func__);
2428 			dump_data((u8 *)tst, ISP1362_BUF_SIZE / 2);
2429 		}
2430 
2431 		for (offset = 0; offset < 256; offset++) {
2432 			int test_size = 0;
2433 
2434 			yield();
2435 
2436 			memset(tst, 0, ISP1362_BUF_SIZE);
2437 			spin_lock_irqsave(&isp1362_hcd->lock, flags);
2438 			isp1362_write_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2439 			isp1362_read_buffer(isp1362_hcd, tst, 0, ISP1362_BUF_SIZE);
2440 			spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2441 			if (memcmp(tst, tst + (ISP1362_BUF_SIZE / (2 * sizeof(*tst))),
2442 				   ISP1362_BUF_SIZE / 2)) {
2443 				pr_err("%s: Failed to clear buffer\n", __func__);
2444 				dump_data((u8 *)tst, ISP1362_BUF_SIZE);
2445 				break;
2446 			}
2447 			spin_lock_irqsave(&isp1362_hcd->lock, flags);
2448 			isp1362_write_buffer(isp1362_hcd, ref, offset * 2, PTD_HEADER_SIZE);
2449 			isp1362_write_buffer(isp1362_hcd, ref + PTD_HEADER_SIZE / sizeof(*ref),
2450 					     offset * 2 + PTD_HEADER_SIZE, test_size);
2451 			isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2452 					    PTD_HEADER_SIZE + test_size);
2453 			spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2454 			if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2455 				dump_data(((u8 *)ref) + offset, PTD_HEADER_SIZE + test_size);
2456 				dump_data((u8 *)tst, PTD_HEADER_SIZE + test_size);
2457 				spin_lock_irqsave(&isp1362_hcd->lock, flags);
2458 				isp1362_read_buffer(isp1362_hcd, tst, offset * 2,
2459 						    PTD_HEADER_SIZE + test_size);
2460 				spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2461 				if (memcmp(ref, tst, PTD_HEADER_SIZE + test_size)) {
2462 					ret = -ENODEV;
2463 					pr_err("%s: memory check with offset %02x failed\n",
2464 					    __func__, offset);
2465 					break;
2466 				}
2467 				pr_warn("%s: memory check with offset %02x ok after second read\n",
2468 					__func__, offset);
2469 			}
2470 		}
2471 		kfree(ref);
2472 	}
2473 	return ret;
2474 }
2475 #endif
2476 
2477 static int isp1362_hc_start(struct usb_hcd *hcd)
2478 {
2479 	int ret;
2480 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2481 	struct isp1362_platform_data *board = isp1362_hcd->board;
2482 	u16 hwcfg;
2483 	u16 chipid;
2484 	unsigned long flags;
2485 
2486 	pr_debug("%s:\n", __func__);
2487 
2488 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2489 	chipid = isp1362_read_reg16(isp1362_hcd, HCCHIPID);
2490 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2491 
2492 	if ((chipid & HCCHIPID_MASK) != HCCHIPID_MAGIC) {
2493 		pr_err("%s: Invalid chip ID %04x\n", __func__, chipid);
2494 		return -ENODEV;
2495 	}
2496 
2497 #ifdef CHIP_BUFFER_TEST
2498 	ret = isp1362_chip_test(isp1362_hcd);
2499 	if (ret)
2500 		return -ENODEV;
2501 #endif
2502 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2503 	/* clear interrupt status and disable all interrupt sources */
2504 	isp1362_write_reg16(isp1362_hcd, HCuPINT, 0xff);
2505 	isp1362_write_reg16(isp1362_hcd, HCuPINTENB, 0);
2506 
2507 	/* HW conf */
2508 	hwcfg = HCHWCFG_INT_ENABLE | HCHWCFG_DBWIDTH(1);
2509 	if (board->sel15Kres)
2510 		hwcfg |= HCHWCFG_PULLDOWN_DS2 |
2511 			((MAX_ROOT_PORTS > 1) ? HCHWCFG_PULLDOWN_DS1 : 0);
2512 	if (board->clknotstop)
2513 		hwcfg |= HCHWCFG_CLKNOTSTOP;
2514 	if (board->oc_enable)
2515 		hwcfg |= HCHWCFG_ANALOG_OC;
2516 	if (board->int_act_high)
2517 		hwcfg |= HCHWCFG_INT_POL;
2518 	if (board->int_edge_triggered)
2519 		hwcfg |= HCHWCFG_INT_TRIGGER;
2520 	if (board->dreq_act_high)
2521 		hwcfg |= HCHWCFG_DREQ_POL;
2522 	if (board->dack_act_high)
2523 		hwcfg |= HCHWCFG_DACK_POL;
2524 	isp1362_write_reg16(isp1362_hcd, HCHWCFG, hwcfg);
2525 	isp1362_show_reg(isp1362_hcd, HCHWCFG);
2526 	isp1362_write_reg16(isp1362_hcd, HCDMACFG, 0);
2527 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2528 
2529 	ret = isp1362_mem_config(hcd);
2530 	if (ret)
2531 		return ret;
2532 
2533 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2534 
2535 	/* Root hub conf */
2536 	isp1362_hcd->rhdesca = 0;
2537 	if (board->no_power_switching)
2538 		isp1362_hcd->rhdesca |= RH_A_NPS;
2539 	if (board->power_switching_mode)
2540 		isp1362_hcd->rhdesca |= RH_A_PSM;
2541 	if (board->potpg)
2542 		isp1362_hcd->rhdesca |= (board->potpg << 24) & RH_A_POTPGT;
2543 	else
2544 		isp1362_hcd->rhdesca |= (25 << 24) & RH_A_POTPGT;
2545 
2546 	isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca & ~RH_A_OCPM);
2547 	isp1362_write_reg32(isp1362_hcd, HCRHDESCA, isp1362_hcd->rhdesca | RH_A_OCPM);
2548 	isp1362_hcd->rhdesca = isp1362_read_reg32(isp1362_hcd, HCRHDESCA);
2549 
2550 	isp1362_hcd->rhdescb = RH_B_PPCM;
2551 	isp1362_write_reg32(isp1362_hcd, HCRHDESCB, isp1362_hcd->rhdescb);
2552 	isp1362_hcd->rhdescb = isp1362_read_reg32(isp1362_hcd, HCRHDESCB);
2553 
2554 	isp1362_read_reg32(isp1362_hcd, HCFMINTVL);
2555 	isp1362_write_reg32(isp1362_hcd, HCFMINTVL, (FSMP(FI) << 16) | FI);
2556 	isp1362_write_reg32(isp1362_hcd, HCLSTHRESH, LSTHRESH);
2557 
2558 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2559 
2560 	isp1362_hcd->hc_control = OHCI_USB_OPER;
2561 	hcd->state = HC_STATE_RUNNING;
2562 
2563 	spin_lock_irqsave(&isp1362_hcd->lock, flags);
2564 	/* Set up interrupts */
2565 	isp1362_hcd->intenb = OHCI_INTR_MIE | OHCI_INTR_RHSC | OHCI_INTR_UE;
2566 	isp1362_hcd->intenb |= OHCI_INTR_RD;
2567 	isp1362_hcd->irqenb = HCuPINT_OPR | HCuPINT_SUSP;
2568 	isp1362_write_reg32(isp1362_hcd, HCINTENB, isp1362_hcd->intenb);
2569 	isp1362_write_reg16(isp1362_hcd, HCuPINTENB, isp1362_hcd->irqenb);
2570 
2571 	/* Go operational */
2572 	isp1362_write_reg32(isp1362_hcd, HCCONTROL, isp1362_hcd->hc_control);
2573 	/* enable global power */
2574 	isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC | RH_HS_DRWE);
2575 
2576 	spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2577 
2578 	return 0;
2579 }
2580 
2581 /*-------------------------------------------------------------------------*/
2582 
2583 static const struct hc_driver isp1362_hc_driver = {
2584 	.description =		hcd_name,
2585 	.product_desc =		"ISP1362 Host Controller",
2586 	.hcd_priv_size =	sizeof(struct isp1362_hcd),
2587 
2588 	.irq =			isp1362_irq,
2589 	.flags =		HCD_USB11 | HCD_MEMORY,
2590 
2591 	.reset =		isp1362_hc_reset,
2592 	.start =		isp1362_hc_start,
2593 	.stop =			isp1362_hc_stop,
2594 
2595 	.urb_enqueue =		isp1362_urb_enqueue,
2596 	.urb_dequeue =		isp1362_urb_dequeue,
2597 	.endpoint_disable =	isp1362_endpoint_disable,
2598 
2599 	.get_frame_number =	isp1362_get_frame,
2600 
2601 	.hub_status_data =	isp1362_hub_status_data,
2602 	.hub_control =		isp1362_hub_control,
2603 	.bus_suspend =		isp1362_bus_suspend,
2604 	.bus_resume =		isp1362_bus_resume,
2605 };
2606 
2607 /*-------------------------------------------------------------------------*/
2608 
2609 static int isp1362_remove(struct platform_device *pdev)
2610 {
2611 	struct usb_hcd *hcd = platform_get_drvdata(pdev);
2612 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2613 
2614 	remove_debug_file(isp1362_hcd);
2615 	DBG(0, "%s: Removing HCD\n", __func__);
2616 	usb_remove_hcd(hcd);
2617 	DBG(0, "%s: put_hcd\n", __func__);
2618 	usb_put_hcd(hcd);
2619 	DBG(0, "%s: Done\n", __func__);
2620 
2621 	return 0;
2622 }
2623 
2624 static int isp1362_probe(struct platform_device *pdev)
2625 {
2626 	struct usb_hcd *hcd;
2627 	struct isp1362_hcd *isp1362_hcd;
2628 	struct resource *data, *irq_res;
2629 	void __iomem *addr_reg;
2630 	void __iomem *data_reg;
2631 	int irq;
2632 	int retval = 0;
2633 	unsigned int irq_flags = 0;
2634 
2635 	if (usb_disabled())
2636 		return -ENODEV;
2637 
2638 	/* basic sanity checks first.  board-specific init logic should
2639 	 * have initialized this the three resources and probably board
2640 	 * specific platform_data.  we don't probe for IRQs, and do only
2641 	 * minimal sanity checking.
2642 	 */
2643 	if (pdev->num_resources < 3)
2644 		return -ENODEV;
2645 
2646 	irq_res = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
2647 	if (!irq_res)
2648 		return -ENODEV;
2649 
2650 	irq = irq_res->start;
2651 
2652 	addr_reg = devm_platform_ioremap_resource(pdev, 1);
2653 	if (IS_ERR(addr_reg))
2654 		return PTR_ERR(addr_reg);
2655 
2656 	data = platform_get_resource(pdev, IORESOURCE_MEM, 0);
2657 	data_reg = devm_ioremap_resource(&pdev->dev, data);
2658 	if (IS_ERR(data_reg))
2659 		return PTR_ERR(data_reg);
2660 
2661 	/* allocate and initialize hcd */
2662 	hcd = usb_create_hcd(&isp1362_hc_driver, &pdev->dev, dev_name(&pdev->dev));
2663 	if (!hcd)
2664 		return -ENOMEM;
2665 
2666 	hcd->rsrc_start = data->start;
2667 	isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2668 	isp1362_hcd->data_reg = data_reg;
2669 	isp1362_hcd->addr_reg = addr_reg;
2670 
2671 	isp1362_hcd->next_statechange = jiffies;
2672 	spin_lock_init(&isp1362_hcd->lock);
2673 	INIT_LIST_HEAD(&isp1362_hcd->async);
2674 	INIT_LIST_HEAD(&isp1362_hcd->periodic);
2675 	INIT_LIST_HEAD(&isp1362_hcd->isoc);
2676 	INIT_LIST_HEAD(&isp1362_hcd->remove_list);
2677 	isp1362_hcd->board = dev_get_platdata(&pdev->dev);
2678 #if USE_PLATFORM_DELAY
2679 	if (!isp1362_hcd->board->delay) {
2680 		dev_err(hcd->self.controller, "No platform delay function given\n");
2681 		retval = -ENODEV;
2682 		goto err;
2683 	}
2684 #endif
2685 
2686 	if (irq_res->flags & IORESOURCE_IRQ_HIGHEDGE)
2687 		irq_flags |= IRQF_TRIGGER_RISING;
2688 	if (irq_res->flags & IORESOURCE_IRQ_LOWEDGE)
2689 		irq_flags |= IRQF_TRIGGER_FALLING;
2690 	if (irq_res->flags & IORESOURCE_IRQ_HIGHLEVEL)
2691 		irq_flags |= IRQF_TRIGGER_HIGH;
2692 	if (irq_res->flags & IORESOURCE_IRQ_LOWLEVEL)
2693 		irq_flags |= IRQF_TRIGGER_LOW;
2694 
2695 	retval = usb_add_hcd(hcd, irq, irq_flags | IRQF_SHARED);
2696 	if (retval != 0)
2697 		goto err;
2698 	device_wakeup_enable(hcd->self.controller);
2699 
2700 	dev_info(&pdev->dev, "%s, irq %d\n", hcd->product_desc, irq);
2701 
2702 	create_debug_file(isp1362_hcd);
2703 
2704 	return 0;
2705 
2706  err:
2707 	usb_put_hcd(hcd);
2708 
2709 	return retval;
2710 }
2711 
2712 #ifdef	CONFIG_PM
2713 static int isp1362_suspend(struct platform_device *pdev, pm_message_t state)
2714 {
2715 	struct usb_hcd *hcd = platform_get_drvdata(pdev);
2716 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2717 	unsigned long flags;
2718 	int retval = 0;
2719 
2720 	DBG(0, "%s: Suspending device\n", __func__);
2721 
2722 	if (state.event == PM_EVENT_FREEZE) {
2723 		DBG(0, "%s: Suspending root hub\n", __func__);
2724 		retval = isp1362_bus_suspend(hcd);
2725 	} else {
2726 		DBG(0, "%s: Suspending RH ports\n", __func__);
2727 		spin_lock_irqsave(&isp1362_hcd->lock, flags);
2728 		isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPS);
2729 		spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2730 	}
2731 	if (retval == 0)
2732 		pdev->dev.power.power_state = state;
2733 	return retval;
2734 }
2735 
2736 static int isp1362_resume(struct platform_device *pdev)
2737 {
2738 	struct usb_hcd *hcd = platform_get_drvdata(pdev);
2739 	struct isp1362_hcd *isp1362_hcd = hcd_to_isp1362_hcd(hcd);
2740 	unsigned long flags;
2741 
2742 	DBG(0, "%s: Resuming\n", __func__);
2743 
2744 	if (pdev->dev.power.power_state.event == PM_EVENT_SUSPEND) {
2745 		DBG(0, "%s: Resume RH ports\n", __func__);
2746 		spin_lock_irqsave(&isp1362_hcd->lock, flags);
2747 		isp1362_write_reg32(isp1362_hcd, HCRHSTATUS, RH_HS_LPSC);
2748 		spin_unlock_irqrestore(&isp1362_hcd->lock, flags);
2749 		return 0;
2750 	}
2751 
2752 	pdev->dev.power.power_state = PMSG_ON;
2753 
2754 	return isp1362_bus_resume(isp1362_hcd_to_hcd(isp1362_hcd));
2755 }
2756 #else
2757 #define	isp1362_suspend	NULL
2758 #define	isp1362_resume	NULL
2759 #endif
2760 
2761 static struct platform_driver isp1362_driver = {
2762 	.probe = isp1362_probe,
2763 	.remove = isp1362_remove,
2764 
2765 	.suspend = isp1362_suspend,
2766 	.resume = isp1362_resume,
2767 	.driver = {
2768 		.name = hcd_name,
2769 	},
2770 };
2771 
2772 module_platform_driver(isp1362_driver);
2773