1 /*
2  * R8A66597 UDC (USB gadget)
3  *
4  * Copyright (C) 2006-2009 Renesas Solutions Corp.
5  *
6  * Author : Yoshihiro Shimoda <yoshihiro.shimoda.uh@renesas.com>
7  *
8  * This program is free software; you can redistribute it and/or modify
9  * it under the terms of the GNU General Public License as published by
10  * the Free Software Foundation; version 2 of the License.
11  */
12 
13 #include <linux/module.h>
14 #include <linux/interrupt.h>
15 #include <linux/delay.h>
16 #include <linux/io.h>
17 #include <linux/platform_device.h>
18 #include <linux/clk.h>
19 #include <linux/err.h>
20 #include <linux/slab.h>
21 #include <linux/dma-mapping.h>
22 
23 #include <linux/usb/ch9.h>
24 #include <linux/usb/gadget.h>
25 
26 #include "r8a66597-udc.h"
27 
28 #define DRIVER_VERSION	"2011-09-26"
29 
30 static const char udc_name[] = "r8a66597_udc";
31 static const char *r8a66597_ep_name[] = {
32 	"ep0", "ep1", "ep2", "ep3", "ep4", "ep5", "ep6", "ep7",
33 	"ep8", "ep9",
34 };
35 
36 static void init_controller(struct r8a66597 *r8a66597);
37 static void disable_controller(struct r8a66597 *r8a66597);
38 static void irq_ep0_write(struct r8a66597_ep *ep, struct r8a66597_request *req);
39 static void irq_packet_write(struct r8a66597_ep *ep,
40 				struct r8a66597_request *req);
41 static int r8a66597_queue(struct usb_ep *_ep, struct usb_request *_req,
42 			gfp_t gfp_flags);
43 
44 static void transfer_complete(struct r8a66597_ep *ep,
45 		struct r8a66597_request *req, int status);
46 
47 /*-------------------------------------------------------------------------*/
48 static inline u16 get_usb_speed(struct r8a66597 *r8a66597)
49 {
50 	return r8a66597_read(r8a66597, DVSTCTR0) & RHST;
51 }
52 
53 static void enable_pipe_irq(struct r8a66597 *r8a66597, u16 pipenum,
54 		unsigned long reg)
55 {
56 	u16 tmp;
57 
58 	tmp = r8a66597_read(r8a66597, INTENB0);
59 	r8a66597_bclr(r8a66597, BEMPE | NRDYE | BRDYE,
60 			INTENB0);
61 	r8a66597_bset(r8a66597, (1 << pipenum), reg);
62 	r8a66597_write(r8a66597, tmp, INTENB0);
63 }
64 
65 static void disable_pipe_irq(struct r8a66597 *r8a66597, u16 pipenum,
66 		unsigned long reg)
67 {
68 	u16 tmp;
69 
70 	tmp = r8a66597_read(r8a66597, INTENB0);
71 	r8a66597_bclr(r8a66597, BEMPE | NRDYE | BRDYE,
72 			INTENB0);
73 	r8a66597_bclr(r8a66597, (1 << pipenum), reg);
74 	r8a66597_write(r8a66597, tmp, INTENB0);
75 }
76 
77 static void r8a66597_usb_connect(struct r8a66597 *r8a66597)
78 {
79 	r8a66597_bset(r8a66597, CTRE, INTENB0);
80 	r8a66597_bset(r8a66597, BEMPE | BRDYE, INTENB0);
81 
82 	r8a66597_bset(r8a66597, DPRPU, SYSCFG0);
83 }
84 
85 static void r8a66597_usb_disconnect(struct r8a66597 *r8a66597)
86 __releases(r8a66597->lock)
87 __acquires(r8a66597->lock)
88 {
89 	r8a66597_bclr(r8a66597, CTRE, INTENB0);
90 	r8a66597_bclr(r8a66597, BEMPE | BRDYE, INTENB0);
91 	r8a66597_bclr(r8a66597, DPRPU, SYSCFG0);
92 
93 	r8a66597->gadget.speed = USB_SPEED_UNKNOWN;
94 	spin_unlock(&r8a66597->lock);
95 	r8a66597->driver->disconnect(&r8a66597->gadget);
96 	spin_lock(&r8a66597->lock);
97 
98 	disable_controller(r8a66597);
99 	init_controller(r8a66597);
100 	r8a66597_bset(r8a66597, VBSE, INTENB0);
101 	INIT_LIST_HEAD(&r8a66597->ep[0].queue);
102 }
103 
104 static inline u16 control_reg_get_pid(struct r8a66597 *r8a66597, u16 pipenum)
105 {
106 	u16 pid = 0;
107 	unsigned long offset;
108 
109 	if (pipenum == 0) {
110 		pid = r8a66597_read(r8a66597, DCPCTR) & PID;
111 	} else if (pipenum < R8A66597_MAX_NUM_PIPE) {
112 		offset = get_pipectr_addr(pipenum);
113 		pid = r8a66597_read(r8a66597, offset) & PID;
114 	} else {
115 		dev_err(r8a66597_to_dev(r8a66597), "unexpect pipe num (%d)\n",
116 			pipenum);
117 	}
118 
119 	return pid;
120 }
121 
122 static inline void control_reg_set_pid(struct r8a66597 *r8a66597, u16 pipenum,
123 		u16 pid)
124 {
125 	unsigned long offset;
126 
127 	if (pipenum == 0) {
128 		r8a66597_mdfy(r8a66597, pid, PID, DCPCTR);
129 	} else if (pipenum < R8A66597_MAX_NUM_PIPE) {
130 		offset = get_pipectr_addr(pipenum);
131 		r8a66597_mdfy(r8a66597, pid, PID, offset);
132 	} else {
133 		dev_err(r8a66597_to_dev(r8a66597), "unexpect pipe num (%d)\n",
134 			pipenum);
135 	}
136 }
137 
138 static inline void pipe_start(struct r8a66597 *r8a66597, u16 pipenum)
139 {
140 	control_reg_set_pid(r8a66597, pipenum, PID_BUF);
141 }
142 
143 static inline void pipe_stop(struct r8a66597 *r8a66597, u16 pipenum)
144 {
145 	control_reg_set_pid(r8a66597, pipenum, PID_NAK);
146 }
147 
148 static inline void pipe_stall(struct r8a66597 *r8a66597, u16 pipenum)
149 {
150 	control_reg_set_pid(r8a66597, pipenum, PID_STALL);
151 }
152 
153 static inline u16 control_reg_get(struct r8a66597 *r8a66597, u16 pipenum)
154 {
155 	u16 ret = 0;
156 	unsigned long offset;
157 
158 	if (pipenum == 0) {
159 		ret = r8a66597_read(r8a66597, DCPCTR);
160 	} else if (pipenum < R8A66597_MAX_NUM_PIPE) {
161 		offset = get_pipectr_addr(pipenum);
162 		ret = r8a66597_read(r8a66597, offset);
163 	} else {
164 		dev_err(r8a66597_to_dev(r8a66597), "unexpect pipe num (%d)\n",
165 			pipenum);
166 	}
167 
168 	return ret;
169 }
170 
171 static inline void control_reg_sqclr(struct r8a66597 *r8a66597, u16 pipenum)
172 {
173 	unsigned long offset;
174 
175 	pipe_stop(r8a66597, pipenum);
176 
177 	if (pipenum == 0) {
178 		r8a66597_bset(r8a66597, SQCLR, DCPCTR);
179 	} else if (pipenum < R8A66597_MAX_NUM_PIPE) {
180 		offset = get_pipectr_addr(pipenum);
181 		r8a66597_bset(r8a66597, SQCLR, offset);
182 	} else {
183 		dev_err(r8a66597_to_dev(r8a66597), "unexpect pipe num (%d)\n",
184 			pipenum);
185 	}
186 }
187 
188 static void control_reg_sqset(struct r8a66597 *r8a66597, u16 pipenum)
189 {
190 	unsigned long offset;
191 
192 	pipe_stop(r8a66597, pipenum);
193 
194 	if (pipenum == 0) {
195 		r8a66597_bset(r8a66597, SQSET, DCPCTR);
196 	} else if (pipenum < R8A66597_MAX_NUM_PIPE) {
197 		offset = get_pipectr_addr(pipenum);
198 		r8a66597_bset(r8a66597, SQSET, offset);
199 	} else {
200 		dev_err(r8a66597_to_dev(r8a66597),
201 			"unexpect pipe num(%d)\n", pipenum);
202 	}
203 }
204 
205 static u16 control_reg_sqmon(struct r8a66597 *r8a66597, u16 pipenum)
206 {
207 	unsigned long offset;
208 
209 	if (pipenum == 0) {
210 		return r8a66597_read(r8a66597, DCPCTR) & SQMON;
211 	} else if (pipenum < R8A66597_MAX_NUM_PIPE) {
212 		offset = get_pipectr_addr(pipenum);
213 		return r8a66597_read(r8a66597, offset) & SQMON;
214 	} else {
215 		dev_err(r8a66597_to_dev(r8a66597),
216 			"unexpect pipe num(%d)\n", pipenum);
217 	}
218 
219 	return 0;
220 }
221 
222 static u16 save_usb_toggle(struct r8a66597 *r8a66597, u16 pipenum)
223 {
224 	return control_reg_sqmon(r8a66597, pipenum);
225 }
226 
227 static void restore_usb_toggle(struct r8a66597 *r8a66597, u16 pipenum,
228 			       u16 toggle)
229 {
230 	if (toggle)
231 		control_reg_sqset(r8a66597, pipenum);
232 	else
233 		control_reg_sqclr(r8a66597, pipenum);
234 }
235 
236 static inline int get_buffer_size(struct r8a66597 *r8a66597, u16 pipenum)
237 {
238 	u16 tmp;
239 	int size;
240 
241 	if (pipenum == 0) {
242 		tmp = r8a66597_read(r8a66597, DCPCFG);
243 		if ((tmp & R8A66597_CNTMD) != 0)
244 			size = 256;
245 		else {
246 			tmp = r8a66597_read(r8a66597, DCPMAXP);
247 			size = tmp & MAXP;
248 		}
249 	} else {
250 		r8a66597_write(r8a66597, pipenum, PIPESEL);
251 		tmp = r8a66597_read(r8a66597, PIPECFG);
252 		if ((tmp & R8A66597_CNTMD) != 0) {
253 			tmp = r8a66597_read(r8a66597, PIPEBUF);
254 			size = ((tmp >> 10) + 1) * 64;
255 		} else {
256 			tmp = r8a66597_read(r8a66597, PIPEMAXP);
257 			size = tmp & MXPS;
258 		}
259 	}
260 
261 	return size;
262 }
263 
264 static inline unsigned short mbw_value(struct r8a66597 *r8a66597)
265 {
266 	if (r8a66597->pdata->on_chip)
267 		return MBW_32;
268 	else
269 		return MBW_16;
270 }
271 
272 static void r8a66597_change_curpipe(struct r8a66597 *r8a66597, u16 pipenum,
273 				    u16 isel, u16 fifosel)
274 {
275 	u16 tmp, mask, loop;
276 	int i = 0;
277 
278 	if (!pipenum) {
279 		mask = ISEL | CURPIPE;
280 		loop = isel;
281 	} else {
282 		mask = CURPIPE;
283 		loop = pipenum;
284 	}
285 	r8a66597_mdfy(r8a66597, loop, mask, fifosel);
286 
287 	do {
288 		tmp = r8a66597_read(r8a66597, fifosel);
289 		if (i++ > 1000000) {
290 			dev_err(r8a66597_to_dev(r8a66597),
291 				"r8a66597: register%x, loop %x "
292 				"is timeout\n", fifosel, loop);
293 			break;
294 		}
295 		ndelay(1);
296 	} while ((tmp & mask) != loop);
297 }
298 
299 static void pipe_change(struct r8a66597 *r8a66597, u16 pipenum)
300 {
301 	struct r8a66597_ep *ep = r8a66597->pipenum2ep[pipenum];
302 
303 	if (ep->use_dma)
304 		r8a66597_bclr(r8a66597, DREQE, ep->fifosel);
305 
306 	r8a66597_mdfy(r8a66597, pipenum, CURPIPE, ep->fifosel);
307 
308 	ndelay(450);
309 
310 	if (r8a66597_is_sudmac(r8a66597) && ep->use_dma)
311 		r8a66597_bclr(r8a66597, mbw_value(r8a66597), ep->fifosel);
312 	else
313 		r8a66597_bset(r8a66597, mbw_value(r8a66597), ep->fifosel);
314 
315 	if (ep->use_dma)
316 		r8a66597_bset(r8a66597, DREQE, ep->fifosel);
317 }
318 
319 static int pipe_buffer_setting(struct r8a66597 *r8a66597,
320 		struct r8a66597_pipe_info *info)
321 {
322 	u16 bufnum = 0, buf_bsize = 0;
323 	u16 pipecfg = 0;
324 
325 	if (info->pipe == 0)
326 		return -EINVAL;
327 
328 	r8a66597_write(r8a66597, info->pipe, PIPESEL);
329 
330 	if (info->dir_in)
331 		pipecfg |= R8A66597_DIR;
332 	pipecfg |= info->type;
333 	pipecfg |= info->epnum;
334 	switch (info->type) {
335 	case R8A66597_INT:
336 		bufnum = 4 + (info->pipe - R8A66597_BASE_PIPENUM_INT);
337 		buf_bsize = 0;
338 		break;
339 	case R8A66597_BULK:
340 		/* isochronous pipes may be used as bulk pipes */
341 		if (info->pipe >= R8A66597_BASE_PIPENUM_BULK)
342 			bufnum = info->pipe - R8A66597_BASE_PIPENUM_BULK;
343 		else
344 			bufnum = info->pipe - R8A66597_BASE_PIPENUM_ISOC;
345 
346 		bufnum = R8A66597_BASE_BUFNUM + (bufnum * 16);
347 		buf_bsize = 7;
348 		pipecfg |= R8A66597_DBLB;
349 		if (!info->dir_in)
350 			pipecfg |= R8A66597_SHTNAK;
351 		break;
352 	case R8A66597_ISO:
353 		bufnum = R8A66597_BASE_BUFNUM +
354 			 (info->pipe - R8A66597_BASE_PIPENUM_ISOC) * 16;
355 		buf_bsize = 7;
356 		break;
357 	}
358 
359 	if (buf_bsize && ((bufnum + 16) >= R8A66597_MAX_BUFNUM)) {
360 		pr_err("r8a66597 pipe memory is insufficient\n");
361 		return -ENOMEM;
362 	}
363 
364 	r8a66597_write(r8a66597, pipecfg, PIPECFG);
365 	r8a66597_write(r8a66597, (buf_bsize << 10) | (bufnum), PIPEBUF);
366 	r8a66597_write(r8a66597, info->maxpacket, PIPEMAXP);
367 	if (info->interval)
368 		info->interval--;
369 	r8a66597_write(r8a66597, info->interval, PIPEPERI);
370 
371 	return 0;
372 }
373 
374 static void pipe_buffer_release(struct r8a66597 *r8a66597,
375 				struct r8a66597_pipe_info *info)
376 {
377 	if (info->pipe == 0)
378 		return;
379 
380 	if (is_bulk_pipe(info->pipe)) {
381 		r8a66597->bulk--;
382 	} else if (is_interrupt_pipe(info->pipe)) {
383 		r8a66597->interrupt--;
384 	} else if (is_isoc_pipe(info->pipe)) {
385 		r8a66597->isochronous--;
386 		if (info->type == R8A66597_BULK)
387 			r8a66597->bulk--;
388 	} else {
389 		dev_err(r8a66597_to_dev(r8a66597),
390 			"ep_release: unexpect pipenum (%d)\n", info->pipe);
391 	}
392 }
393 
394 static void pipe_initialize(struct r8a66597_ep *ep)
395 {
396 	struct r8a66597 *r8a66597 = ep->r8a66597;
397 
398 	r8a66597_mdfy(r8a66597, 0, CURPIPE, ep->fifosel);
399 
400 	r8a66597_write(r8a66597, ACLRM, ep->pipectr);
401 	r8a66597_write(r8a66597, 0, ep->pipectr);
402 	r8a66597_write(r8a66597, SQCLR, ep->pipectr);
403 	if (ep->use_dma) {
404 		r8a66597_mdfy(r8a66597, ep->pipenum, CURPIPE, ep->fifosel);
405 
406 		ndelay(450);
407 
408 		r8a66597_bset(r8a66597, mbw_value(r8a66597), ep->fifosel);
409 	}
410 }
411 
412 static void r8a66597_ep_setting(struct r8a66597 *r8a66597,
413 				struct r8a66597_ep *ep,
414 				const struct usb_endpoint_descriptor *desc,
415 				u16 pipenum, int dma)
416 {
417 	ep->use_dma = 0;
418 	ep->fifoaddr = CFIFO;
419 	ep->fifosel = CFIFOSEL;
420 	ep->fifoctr = CFIFOCTR;
421 
422 	ep->pipectr = get_pipectr_addr(pipenum);
423 	if (is_bulk_pipe(pipenum) || is_isoc_pipe(pipenum)) {
424 		ep->pipetre = get_pipetre_addr(pipenum);
425 		ep->pipetrn = get_pipetrn_addr(pipenum);
426 	} else {
427 		ep->pipetre = 0;
428 		ep->pipetrn = 0;
429 	}
430 	ep->pipenum = pipenum;
431 	ep->ep.maxpacket = usb_endpoint_maxp(desc);
432 	r8a66597->pipenum2ep[pipenum] = ep;
433 	r8a66597->epaddr2ep[usb_endpoint_num(desc)]
434 		= ep;
435 	INIT_LIST_HEAD(&ep->queue);
436 }
437 
438 static void r8a66597_ep_release(struct r8a66597_ep *ep)
439 {
440 	struct r8a66597 *r8a66597 = ep->r8a66597;
441 	u16 pipenum = ep->pipenum;
442 
443 	if (pipenum == 0)
444 		return;
445 
446 	if (ep->use_dma)
447 		r8a66597->num_dma--;
448 	ep->pipenum = 0;
449 	ep->busy = 0;
450 	ep->use_dma = 0;
451 }
452 
453 static int alloc_pipe_config(struct r8a66597_ep *ep,
454 		const struct usb_endpoint_descriptor *desc)
455 {
456 	struct r8a66597 *r8a66597 = ep->r8a66597;
457 	struct r8a66597_pipe_info info;
458 	int dma = 0;
459 	unsigned char *counter;
460 	int ret;
461 
462 	ep->ep.desc = desc;
463 
464 	if (ep->pipenum)	/* already allocated pipe  */
465 		return 0;
466 
467 	switch (usb_endpoint_type(desc)) {
468 	case USB_ENDPOINT_XFER_BULK:
469 		if (r8a66597->bulk >= R8A66597_MAX_NUM_BULK) {
470 			if (r8a66597->isochronous >= R8A66597_MAX_NUM_ISOC) {
471 				dev_err(r8a66597_to_dev(r8a66597),
472 					"bulk pipe is insufficient\n");
473 				return -ENODEV;
474 			} else {
475 				info.pipe = R8A66597_BASE_PIPENUM_ISOC
476 						+ r8a66597->isochronous;
477 				counter = &r8a66597->isochronous;
478 			}
479 		} else {
480 			info.pipe = R8A66597_BASE_PIPENUM_BULK + r8a66597->bulk;
481 			counter = &r8a66597->bulk;
482 		}
483 		info.type = R8A66597_BULK;
484 		dma = 1;
485 		break;
486 	case USB_ENDPOINT_XFER_INT:
487 		if (r8a66597->interrupt >= R8A66597_MAX_NUM_INT) {
488 			dev_err(r8a66597_to_dev(r8a66597),
489 				"interrupt pipe is insufficient\n");
490 			return -ENODEV;
491 		}
492 		info.pipe = R8A66597_BASE_PIPENUM_INT + r8a66597->interrupt;
493 		info.type = R8A66597_INT;
494 		counter = &r8a66597->interrupt;
495 		break;
496 	case USB_ENDPOINT_XFER_ISOC:
497 		if (r8a66597->isochronous >= R8A66597_MAX_NUM_ISOC) {
498 			dev_err(r8a66597_to_dev(r8a66597),
499 				"isochronous pipe is insufficient\n");
500 			return -ENODEV;
501 		}
502 		info.pipe = R8A66597_BASE_PIPENUM_ISOC + r8a66597->isochronous;
503 		info.type = R8A66597_ISO;
504 		counter = &r8a66597->isochronous;
505 		break;
506 	default:
507 		dev_err(r8a66597_to_dev(r8a66597), "unexpect xfer type\n");
508 		return -EINVAL;
509 	}
510 	ep->type = info.type;
511 
512 	info.epnum = usb_endpoint_num(desc);
513 	info.maxpacket = usb_endpoint_maxp(desc);
514 	info.interval = desc->bInterval;
515 	if (desc->bEndpointAddress & USB_ENDPOINT_DIR_MASK)
516 		info.dir_in = 1;
517 	else
518 		info.dir_in = 0;
519 
520 	ret = pipe_buffer_setting(r8a66597, &info);
521 	if (ret < 0) {
522 		dev_err(r8a66597_to_dev(r8a66597),
523 			"pipe_buffer_setting fail\n");
524 		return ret;
525 	}
526 
527 	(*counter)++;
528 	if ((counter == &r8a66597->isochronous) && info.type == R8A66597_BULK)
529 		r8a66597->bulk++;
530 
531 	r8a66597_ep_setting(r8a66597, ep, desc, info.pipe, dma);
532 	pipe_initialize(ep);
533 
534 	return 0;
535 }
536 
537 static int free_pipe_config(struct r8a66597_ep *ep)
538 {
539 	struct r8a66597 *r8a66597 = ep->r8a66597;
540 	struct r8a66597_pipe_info info;
541 
542 	info.pipe = ep->pipenum;
543 	info.type = ep->type;
544 	pipe_buffer_release(r8a66597, &info);
545 	r8a66597_ep_release(ep);
546 
547 	return 0;
548 }
549 
550 /*-------------------------------------------------------------------------*/
551 static void pipe_irq_enable(struct r8a66597 *r8a66597, u16 pipenum)
552 {
553 	enable_irq_ready(r8a66597, pipenum);
554 	enable_irq_nrdy(r8a66597, pipenum);
555 }
556 
557 static void pipe_irq_disable(struct r8a66597 *r8a66597, u16 pipenum)
558 {
559 	disable_irq_ready(r8a66597, pipenum);
560 	disable_irq_nrdy(r8a66597, pipenum);
561 }
562 
563 /* if complete is true, gadget driver complete function is not call */
564 static void control_end(struct r8a66597 *r8a66597, unsigned ccpl)
565 {
566 	r8a66597->ep[0].internal_ccpl = ccpl;
567 	pipe_start(r8a66597, 0);
568 	r8a66597_bset(r8a66597, CCPL, DCPCTR);
569 }
570 
571 static void start_ep0_write(struct r8a66597_ep *ep,
572 				struct r8a66597_request *req)
573 {
574 	struct r8a66597 *r8a66597 = ep->r8a66597;
575 
576 	pipe_change(r8a66597, ep->pipenum);
577 	r8a66597_mdfy(r8a66597, ISEL, (ISEL | CURPIPE), CFIFOSEL);
578 	r8a66597_write(r8a66597, BCLR, ep->fifoctr);
579 	if (req->req.length == 0) {
580 		r8a66597_bset(r8a66597, BVAL, ep->fifoctr);
581 		pipe_start(r8a66597, 0);
582 		transfer_complete(ep, req, 0);
583 	} else {
584 		r8a66597_write(r8a66597, ~BEMP0, BEMPSTS);
585 		irq_ep0_write(ep, req);
586 	}
587 }
588 
589 static void disable_fifosel(struct r8a66597 *r8a66597, u16 pipenum,
590 			    u16 fifosel)
591 {
592 	u16 tmp;
593 
594 	tmp = r8a66597_read(r8a66597, fifosel) & CURPIPE;
595 	if (tmp == pipenum)
596 		r8a66597_change_curpipe(r8a66597, 0, 0, fifosel);
597 }
598 
599 static void change_bfre_mode(struct r8a66597 *r8a66597, u16 pipenum,
600 			     int enable)
601 {
602 	struct r8a66597_ep *ep = r8a66597->pipenum2ep[pipenum];
603 	u16 tmp, toggle;
604 
605 	/* check current BFRE bit */
606 	r8a66597_write(r8a66597, pipenum, PIPESEL);
607 	tmp = r8a66597_read(r8a66597, PIPECFG) & R8A66597_BFRE;
608 	if ((enable && tmp) || (!enable && !tmp))
609 		return;
610 
611 	/* change BFRE bit */
612 	pipe_stop(r8a66597, pipenum);
613 	disable_fifosel(r8a66597, pipenum, CFIFOSEL);
614 	disable_fifosel(r8a66597, pipenum, D0FIFOSEL);
615 	disable_fifosel(r8a66597, pipenum, D1FIFOSEL);
616 
617 	toggle = save_usb_toggle(r8a66597, pipenum);
618 
619 	r8a66597_write(r8a66597, pipenum, PIPESEL);
620 	if (enable)
621 		r8a66597_bset(r8a66597, R8A66597_BFRE, PIPECFG);
622 	else
623 		r8a66597_bclr(r8a66597, R8A66597_BFRE, PIPECFG);
624 
625 	/* initialize for internal BFRE flag */
626 	r8a66597_bset(r8a66597, ACLRM, ep->pipectr);
627 	r8a66597_bclr(r8a66597, ACLRM, ep->pipectr);
628 
629 	restore_usb_toggle(r8a66597, pipenum, toggle);
630 }
631 
632 static int sudmac_alloc_channel(struct r8a66597 *r8a66597,
633 				struct r8a66597_ep *ep,
634 				struct r8a66597_request *req)
635 {
636 	struct r8a66597_dma *dma;
637 
638 	if (!r8a66597_is_sudmac(r8a66597))
639 		return -ENODEV;
640 
641 	/* Check transfer type */
642 	if (!is_bulk_pipe(ep->pipenum))
643 		return -EIO;
644 
645 	if (r8a66597->dma.used)
646 		return -EBUSY;
647 
648 	/* set SUDMAC parameters */
649 	dma = &r8a66597->dma;
650 	dma->used = 1;
651 	if (ep->ep.desc->bEndpointAddress & USB_DIR_IN) {
652 		dma->dir = 1;
653 	} else {
654 		dma->dir = 0;
655 		change_bfre_mode(r8a66597, ep->pipenum, 1);
656 	}
657 
658 	/* set r8a66597_ep paramters */
659 	ep->use_dma = 1;
660 	ep->dma = dma;
661 	ep->fifoaddr = D0FIFO;
662 	ep->fifosel = D0FIFOSEL;
663 	ep->fifoctr = D0FIFOCTR;
664 
665 	/* dma mapping */
666 	return usb_gadget_map_request(&r8a66597->gadget, &req->req, dma->dir);
667 }
668 
669 static void sudmac_free_channel(struct r8a66597 *r8a66597,
670 				struct r8a66597_ep *ep,
671 				struct r8a66597_request *req)
672 {
673 	if (!r8a66597_is_sudmac(r8a66597))
674 		return;
675 
676 	usb_gadget_unmap_request(&r8a66597->gadget, &req->req, ep->dma->dir);
677 
678 	r8a66597_bclr(r8a66597, DREQE, ep->fifosel);
679 	r8a66597_change_curpipe(r8a66597, 0, 0, ep->fifosel);
680 
681 	ep->dma->used = 0;
682 	ep->use_dma = 0;
683 	ep->fifoaddr = CFIFO;
684 	ep->fifosel = CFIFOSEL;
685 	ep->fifoctr = CFIFOCTR;
686 }
687 
688 static void sudmac_start(struct r8a66597 *r8a66597, struct r8a66597_ep *ep,
689 			 struct r8a66597_request *req)
690 {
691 	BUG_ON(req->req.length == 0);
692 
693 	r8a66597_sudmac_write(r8a66597, LBA_WAIT, CH0CFG);
694 	r8a66597_sudmac_write(r8a66597, req->req.dma, CH0BA);
695 	r8a66597_sudmac_write(r8a66597, req->req.length, CH0BBC);
696 	r8a66597_sudmac_write(r8a66597, CH0ENDE, DINTCTRL);
697 
698 	r8a66597_sudmac_write(r8a66597, DEN, CH0DEN);
699 }
700 
701 static void start_packet_write(struct r8a66597_ep *ep,
702 				struct r8a66597_request *req)
703 {
704 	struct r8a66597 *r8a66597 = ep->r8a66597;
705 	u16 tmp;
706 
707 	pipe_change(r8a66597, ep->pipenum);
708 	disable_irq_empty(r8a66597, ep->pipenum);
709 	pipe_start(r8a66597, ep->pipenum);
710 
711 	if (req->req.length == 0) {
712 		transfer_complete(ep, req, 0);
713 	} else {
714 		r8a66597_write(r8a66597, ~(1 << ep->pipenum), BRDYSTS);
715 		if (sudmac_alloc_channel(r8a66597, ep, req) < 0) {
716 			/* PIO mode */
717 			pipe_change(r8a66597, ep->pipenum);
718 			disable_irq_empty(r8a66597, ep->pipenum);
719 			pipe_start(r8a66597, ep->pipenum);
720 			tmp = r8a66597_read(r8a66597, ep->fifoctr);
721 			if (unlikely((tmp & FRDY) == 0))
722 				pipe_irq_enable(r8a66597, ep->pipenum);
723 			else
724 				irq_packet_write(ep, req);
725 		} else {
726 			/* DMA mode */
727 			pipe_change(r8a66597, ep->pipenum);
728 			disable_irq_nrdy(r8a66597, ep->pipenum);
729 			pipe_start(r8a66597, ep->pipenum);
730 			enable_irq_nrdy(r8a66597, ep->pipenum);
731 			sudmac_start(r8a66597, ep, req);
732 		}
733 	}
734 }
735 
736 static void start_packet_read(struct r8a66597_ep *ep,
737 				struct r8a66597_request *req)
738 {
739 	struct r8a66597 *r8a66597 = ep->r8a66597;
740 	u16 pipenum = ep->pipenum;
741 
742 	if (ep->pipenum == 0) {
743 		r8a66597_mdfy(r8a66597, 0, (ISEL | CURPIPE), CFIFOSEL);
744 		r8a66597_write(r8a66597, BCLR, ep->fifoctr);
745 		pipe_start(r8a66597, pipenum);
746 		pipe_irq_enable(r8a66597, pipenum);
747 	} else {
748 		pipe_stop(r8a66597, pipenum);
749 		if (ep->pipetre) {
750 			enable_irq_nrdy(r8a66597, pipenum);
751 			r8a66597_write(r8a66597, TRCLR, ep->pipetre);
752 			r8a66597_write(r8a66597,
753 				DIV_ROUND_UP(req->req.length, ep->ep.maxpacket),
754 				ep->pipetrn);
755 			r8a66597_bset(r8a66597, TRENB, ep->pipetre);
756 		}
757 
758 		if (sudmac_alloc_channel(r8a66597, ep, req) < 0) {
759 			/* PIO mode */
760 			change_bfre_mode(r8a66597, ep->pipenum, 0);
761 			pipe_start(r8a66597, pipenum);	/* trigger once */
762 			pipe_irq_enable(r8a66597, pipenum);
763 		} else {
764 			pipe_change(r8a66597, pipenum);
765 			sudmac_start(r8a66597, ep, req);
766 			pipe_start(r8a66597, pipenum);	/* trigger once */
767 		}
768 	}
769 }
770 
771 static void start_packet(struct r8a66597_ep *ep, struct r8a66597_request *req)
772 {
773 	if (ep->ep.desc->bEndpointAddress & USB_DIR_IN)
774 		start_packet_write(ep, req);
775 	else
776 		start_packet_read(ep, req);
777 }
778 
779 static void start_ep0(struct r8a66597_ep *ep, struct r8a66597_request *req)
780 {
781 	u16 ctsq;
782 
783 	ctsq = r8a66597_read(ep->r8a66597, INTSTS0) & CTSQ;
784 
785 	switch (ctsq) {
786 	case CS_RDDS:
787 		start_ep0_write(ep, req);
788 		break;
789 	case CS_WRDS:
790 		start_packet_read(ep, req);
791 		break;
792 
793 	case CS_WRND:
794 		control_end(ep->r8a66597, 0);
795 		break;
796 	default:
797 		dev_err(r8a66597_to_dev(ep->r8a66597),
798 			"start_ep0: unexpect ctsq(%x)\n", ctsq);
799 		break;
800 	}
801 }
802 
803 static void init_controller(struct r8a66597 *r8a66597)
804 {
805 	u16 vif = r8a66597->pdata->vif ? LDRV : 0;
806 	u16 irq_sense = r8a66597->irq_sense_low ? INTL : 0;
807 	u16 endian = r8a66597->pdata->endian ? BIGEND : 0;
808 
809 	if (r8a66597->pdata->on_chip) {
810 		if (r8a66597->pdata->buswait)
811 			r8a66597_write(r8a66597, r8a66597->pdata->buswait,
812 					SYSCFG1);
813 		else
814 			r8a66597_write(r8a66597, 0x0f, SYSCFG1);
815 		r8a66597_bset(r8a66597, HSE, SYSCFG0);
816 
817 		r8a66597_bclr(r8a66597, USBE, SYSCFG0);
818 		r8a66597_bclr(r8a66597, DPRPU, SYSCFG0);
819 		r8a66597_bset(r8a66597, USBE, SYSCFG0);
820 
821 		r8a66597_bset(r8a66597, SCKE, SYSCFG0);
822 
823 		r8a66597_bset(r8a66597, irq_sense, INTENB1);
824 		r8a66597_write(r8a66597, BURST | CPU_ADR_RD_WR,
825 				DMA0CFG);
826 	} else {
827 		r8a66597_bset(r8a66597, vif | endian, PINCFG);
828 		r8a66597_bset(r8a66597, HSE, SYSCFG0);		/* High spd */
829 		r8a66597_mdfy(r8a66597, get_xtal_from_pdata(r8a66597->pdata),
830 				XTAL, SYSCFG0);
831 
832 		r8a66597_bclr(r8a66597, USBE, SYSCFG0);
833 		r8a66597_bclr(r8a66597, DPRPU, SYSCFG0);
834 		r8a66597_bset(r8a66597, USBE, SYSCFG0);
835 
836 		r8a66597_bset(r8a66597, XCKE, SYSCFG0);
837 
838 		msleep(3);
839 
840 		r8a66597_bset(r8a66597, PLLC, SYSCFG0);
841 
842 		msleep(1);
843 
844 		r8a66597_bset(r8a66597, SCKE, SYSCFG0);
845 
846 		r8a66597_bset(r8a66597, irq_sense, INTENB1);
847 		r8a66597_write(r8a66597, BURST | CPU_ADR_RD_WR,
848 			       DMA0CFG);
849 	}
850 }
851 
852 static void disable_controller(struct r8a66597 *r8a66597)
853 {
854 	if (r8a66597->pdata->on_chip) {
855 		r8a66597_bset(r8a66597, SCKE, SYSCFG0);
856 		r8a66597_bclr(r8a66597, UTST, TESTMODE);
857 
858 		/* disable interrupts */
859 		r8a66597_write(r8a66597, 0, INTENB0);
860 		r8a66597_write(r8a66597, 0, INTENB1);
861 		r8a66597_write(r8a66597, 0, BRDYENB);
862 		r8a66597_write(r8a66597, 0, BEMPENB);
863 		r8a66597_write(r8a66597, 0, NRDYENB);
864 
865 		/* clear status */
866 		r8a66597_write(r8a66597, 0, BRDYSTS);
867 		r8a66597_write(r8a66597, 0, NRDYSTS);
868 		r8a66597_write(r8a66597, 0, BEMPSTS);
869 
870 		r8a66597_bclr(r8a66597, USBE, SYSCFG0);
871 		r8a66597_bclr(r8a66597, SCKE, SYSCFG0);
872 
873 	} else {
874 		r8a66597_bclr(r8a66597, UTST, TESTMODE);
875 		r8a66597_bclr(r8a66597, SCKE, SYSCFG0);
876 		udelay(1);
877 		r8a66597_bclr(r8a66597, PLLC, SYSCFG0);
878 		udelay(1);
879 		udelay(1);
880 		r8a66597_bclr(r8a66597, XCKE, SYSCFG0);
881 	}
882 }
883 
884 static void r8a66597_start_xclock(struct r8a66597 *r8a66597)
885 {
886 	u16 tmp;
887 
888 	if (!r8a66597->pdata->on_chip) {
889 		tmp = r8a66597_read(r8a66597, SYSCFG0);
890 		if (!(tmp & XCKE))
891 			r8a66597_bset(r8a66597, XCKE, SYSCFG0);
892 	}
893 }
894 
895 static struct r8a66597_request *get_request_from_ep(struct r8a66597_ep *ep)
896 {
897 	return list_entry(ep->queue.next, struct r8a66597_request, queue);
898 }
899 
900 /*-------------------------------------------------------------------------*/
901 static void transfer_complete(struct r8a66597_ep *ep,
902 		struct r8a66597_request *req, int status)
903 __releases(r8a66597->lock)
904 __acquires(r8a66597->lock)
905 {
906 	int restart = 0;
907 
908 	if (unlikely(ep->pipenum == 0)) {
909 		if (ep->internal_ccpl) {
910 			ep->internal_ccpl = 0;
911 			return;
912 		}
913 	}
914 
915 	list_del_init(&req->queue);
916 	if (ep->r8a66597->gadget.speed == USB_SPEED_UNKNOWN)
917 		req->req.status = -ESHUTDOWN;
918 	else
919 		req->req.status = status;
920 
921 	if (!list_empty(&ep->queue))
922 		restart = 1;
923 
924 	if (ep->use_dma)
925 		sudmac_free_channel(ep->r8a66597, ep, req);
926 
927 	spin_unlock(&ep->r8a66597->lock);
928 	usb_gadget_giveback_request(&ep->ep, &req->req);
929 	spin_lock(&ep->r8a66597->lock);
930 
931 	if (restart) {
932 		req = get_request_from_ep(ep);
933 		if (ep->ep.desc)
934 			start_packet(ep, req);
935 	}
936 }
937 
938 static void irq_ep0_write(struct r8a66597_ep *ep, struct r8a66597_request *req)
939 {
940 	int i;
941 	u16 tmp;
942 	unsigned bufsize;
943 	size_t size;
944 	void *buf;
945 	u16 pipenum = ep->pipenum;
946 	struct r8a66597 *r8a66597 = ep->r8a66597;
947 
948 	pipe_change(r8a66597, pipenum);
949 	r8a66597_bset(r8a66597, ISEL, ep->fifosel);
950 
951 	i = 0;
952 	do {
953 		tmp = r8a66597_read(r8a66597, ep->fifoctr);
954 		if (i++ > 100000) {
955 			dev_err(r8a66597_to_dev(r8a66597),
956 				"pipe0 is busy. maybe cpu i/o bus "
957 				"conflict. please power off this controller.");
958 			return;
959 		}
960 		ndelay(1);
961 	} while ((tmp & FRDY) == 0);
962 
963 	/* prepare parameters */
964 	bufsize = get_buffer_size(r8a66597, pipenum);
965 	buf = req->req.buf + req->req.actual;
966 	size = min(bufsize, req->req.length - req->req.actual);
967 
968 	/* write fifo */
969 	if (req->req.buf) {
970 		if (size > 0)
971 			r8a66597_write_fifo(r8a66597, ep, buf, size);
972 		if ((size == 0) || ((size % ep->ep.maxpacket) != 0))
973 			r8a66597_bset(r8a66597, BVAL, ep->fifoctr);
974 	}
975 
976 	/* update parameters */
977 	req->req.actual += size;
978 
979 	/* check transfer finish */
980 	if ((!req->req.zero && (req->req.actual == req->req.length))
981 			|| (size % ep->ep.maxpacket)
982 			|| (size == 0)) {
983 		disable_irq_ready(r8a66597, pipenum);
984 		disable_irq_empty(r8a66597, pipenum);
985 	} else {
986 		disable_irq_ready(r8a66597, pipenum);
987 		enable_irq_empty(r8a66597, pipenum);
988 	}
989 	pipe_start(r8a66597, pipenum);
990 }
991 
992 static void irq_packet_write(struct r8a66597_ep *ep,
993 				struct r8a66597_request *req)
994 {
995 	u16 tmp;
996 	unsigned bufsize;
997 	size_t size;
998 	void *buf;
999 	u16 pipenum = ep->pipenum;
1000 	struct r8a66597 *r8a66597 = ep->r8a66597;
1001 
1002 	pipe_change(r8a66597, pipenum);
1003 	tmp = r8a66597_read(r8a66597, ep->fifoctr);
1004 	if (unlikely((tmp & FRDY) == 0)) {
1005 		pipe_stop(r8a66597, pipenum);
1006 		pipe_irq_disable(r8a66597, pipenum);
1007 		dev_err(r8a66597_to_dev(r8a66597),
1008 			"write fifo not ready. pipnum=%d\n", pipenum);
1009 		return;
1010 	}
1011 
1012 	/* prepare parameters */
1013 	bufsize = get_buffer_size(r8a66597, pipenum);
1014 	buf = req->req.buf + req->req.actual;
1015 	size = min(bufsize, req->req.length - req->req.actual);
1016 
1017 	/* write fifo */
1018 	if (req->req.buf) {
1019 		r8a66597_write_fifo(r8a66597, ep, buf, size);
1020 		if ((size == 0)
1021 				|| ((size % ep->ep.maxpacket) != 0)
1022 				|| ((bufsize != ep->ep.maxpacket)
1023 					&& (bufsize > size)))
1024 			r8a66597_bset(r8a66597, BVAL, ep->fifoctr);
1025 	}
1026 
1027 	/* update parameters */
1028 	req->req.actual += size;
1029 	/* check transfer finish */
1030 	if ((!req->req.zero && (req->req.actual == req->req.length))
1031 			|| (size % ep->ep.maxpacket)
1032 			|| (size == 0)) {
1033 		disable_irq_ready(r8a66597, pipenum);
1034 		enable_irq_empty(r8a66597, pipenum);
1035 	} else {
1036 		disable_irq_empty(r8a66597, pipenum);
1037 		pipe_irq_enable(r8a66597, pipenum);
1038 	}
1039 }
1040 
1041 static void irq_packet_read(struct r8a66597_ep *ep,
1042 				struct r8a66597_request *req)
1043 {
1044 	u16 tmp;
1045 	int rcv_len, bufsize, req_len;
1046 	int size;
1047 	void *buf;
1048 	u16 pipenum = ep->pipenum;
1049 	struct r8a66597 *r8a66597 = ep->r8a66597;
1050 	int finish = 0;
1051 
1052 	pipe_change(r8a66597, pipenum);
1053 	tmp = r8a66597_read(r8a66597, ep->fifoctr);
1054 	if (unlikely((tmp & FRDY) == 0)) {
1055 		req->req.status = -EPIPE;
1056 		pipe_stop(r8a66597, pipenum);
1057 		pipe_irq_disable(r8a66597, pipenum);
1058 		dev_err(r8a66597_to_dev(r8a66597), "read fifo not ready");
1059 		return;
1060 	}
1061 
1062 	/* prepare parameters */
1063 	rcv_len = tmp & DTLN;
1064 	bufsize = get_buffer_size(r8a66597, pipenum);
1065 
1066 	buf = req->req.buf + req->req.actual;
1067 	req_len = req->req.length - req->req.actual;
1068 	if (rcv_len < bufsize)
1069 		size = min(rcv_len, req_len);
1070 	else
1071 		size = min(bufsize, req_len);
1072 
1073 	/* update parameters */
1074 	req->req.actual += size;
1075 
1076 	/* check transfer finish */
1077 	if ((!req->req.zero && (req->req.actual == req->req.length))
1078 			|| (size % ep->ep.maxpacket)
1079 			|| (size == 0)) {
1080 		pipe_stop(r8a66597, pipenum);
1081 		pipe_irq_disable(r8a66597, pipenum);
1082 		finish = 1;
1083 	}
1084 
1085 	/* read fifo */
1086 	if (req->req.buf) {
1087 		if (size == 0)
1088 			r8a66597_write(r8a66597, BCLR, ep->fifoctr);
1089 		else
1090 			r8a66597_read_fifo(r8a66597, ep->fifoaddr, buf, size);
1091 
1092 	}
1093 
1094 	if ((ep->pipenum != 0) && finish)
1095 		transfer_complete(ep, req, 0);
1096 }
1097 
1098 static void irq_pipe_ready(struct r8a66597 *r8a66597, u16 status, u16 enb)
1099 {
1100 	u16 check;
1101 	u16 pipenum;
1102 	struct r8a66597_ep *ep;
1103 	struct r8a66597_request *req;
1104 
1105 	if ((status & BRDY0) && (enb & BRDY0)) {
1106 		r8a66597_write(r8a66597, ~BRDY0, BRDYSTS);
1107 		r8a66597_mdfy(r8a66597, 0, CURPIPE, CFIFOSEL);
1108 
1109 		ep = &r8a66597->ep[0];
1110 		req = get_request_from_ep(ep);
1111 		irq_packet_read(ep, req);
1112 	} else {
1113 		for (pipenum = 1; pipenum < R8A66597_MAX_NUM_PIPE; pipenum++) {
1114 			check = 1 << pipenum;
1115 			if ((status & check) && (enb & check)) {
1116 				r8a66597_write(r8a66597, ~check, BRDYSTS);
1117 				ep = r8a66597->pipenum2ep[pipenum];
1118 				req = get_request_from_ep(ep);
1119 				if (ep->ep.desc->bEndpointAddress & USB_DIR_IN)
1120 					irq_packet_write(ep, req);
1121 				else
1122 					irq_packet_read(ep, req);
1123 			}
1124 		}
1125 	}
1126 }
1127 
1128 static void irq_pipe_empty(struct r8a66597 *r8a66597, u16 status, u16 enb)
1129 {
1130 	u16 tmp;
1131 	u16 check;
1132 	u16 pipenum;
1133 	struct r8a66597_ep *ep;
1134 	struct r8a66597_request *req;
1135 
1136 	if ((status & BEMP0) && (enb & BEMP0)) {
1137 		r8a66597_write(r8a66597, ~BEMP0, BEMPSTS);
1138 
1139 		ep = &r8a66597->ep[0];
1140 		req = get_request_from_ep(ep);
1141 		irq_ep0_write(ep, req);
1142 	} else {
1143 		for (pipenum = 1; pipenum < R8A66597_MAX_NUM_PIPE; pipenum++) {
1144 			check = 1 << pipenum;
1145 			if ((status & check) && (enb & check)) {
1146 				r8a66597_write(r8a66597, ~check, BEMPSTS);
1147 				tmp = control_reg_get(r8a66597, pipenum);
1148 				if ((tmp & INBUFM) == 0) {
1149 					disable_irq_empty(r8a66597, pipenum);
1150 					pipe_irq_disable(r8a66597, pipenum);
1151 					pipe_stop(r8a66597, pipenum);
1152 					ep = r8a66597->pipenum2ep[pipenum];
1153 					req = get_request_from_ep(ep);
1154 					if (!list_empty(&ep->queue))
1155 						transfer_complete(ep, req, 0);
1156 				}
1157 			}
1158 		}
1159 	}
1160 }
1161 
1162 static void get_status(struct r8a66597 *r8a66597, struct usb_ctrlrequest *ctrl)
1163 __releases(r8a66597->lock)
1164 __acquires(r8a66597->lock)
1165 {
1166 	struct r8a66597_ep *ep;
1167 	u16 pid;
1168 	u16 status = 0;
1169 	u16 w_index = le16_to_cpu(ctrl->wIndex);
1170 
1171 	switch (ctrl->bRequestType & USB_RECIP_MASK) {
1172 	case USB_RECIP_DEVICE:
1173 		status = r8a66597->device_status;
1174 		break;
1175 	case USB_RECIP_INTERFACE:
1176 		status = 0;
1177 		break;
1178 	case USB_RECIP_ENDPOINT:
1179 		ep = r8a66597->epaddr2ep[w_index & USB_ENDPOINT_NUMBER_MASK];
1180 		pid = control_reg_get_pid(r8a66597, ep->pipenum);
1181 		if (pid == PID_STALL)
1182 			status = 1 << USB_ENDPOINT_HALT;
1183 		else
1184 			status = 0;
1185 		break;
1186 	default:
1187 		pipe_stall(r8a66597, 0);
1188 		return;		/* exit */
1189 	}
1190 
1191 	r8a66597->ep0_data = cpu_to_le16(status);
1192 	r8a66597->ep0_req->buf = &r8a66597->ep0_data;
1193 	r8a66597->ep0_req->length = 2;
1194 	/* AV: what happens if we get called again before that gets through? */
1195 	spin_unlock(&r8a66597->lock);
1196 	r8a66597_queue(r8a66597->gadget.ep0, r8a66597->ep0_req, GFP_KERNEL);
1197 	spin_lock(&r8a66597->lock);
1198 }
1199 
1200 static void clear_feature(struct r8a66597 *r8a66597,
1201 				struct usb_ctrlrequest *ctrl)
1202 {
1203 	switch (ctrl->bRequestType & USB_RECIP_MASK) {
1204 	case USB_RECIP_DEVICE:
1205 		control_end(r8a66597, 1);
1206 		break;
1207 	case USB_RECIP_INTERFACE:
1208 		control_end(r8a66597, 1);
1209 		break;
1210 	case USB_RECIP_ENDPOINT: {
1211 		struct r8a66597_ep *ep;
1212 		struct r8a66597_request *req;
1213 		u16 w_index = le16_to_cpu(ctrl->wIndex);
1214 
1215 		ep = r8a66597->epaddr2ep[w_index & USB_ENDPOINT_NUMBER_MASK];
1216 		if (!ep->wedge) {
1217 			pipe_stop(r8a66597, ep->pipenum);
1218 			control_reg_sqclr(r8a66597, ep->pipenum);
1219 			spin_unlock(&r8a66597->lock);
1220 			usb_ep_clear_halt(&ep->ep);
1221 			spin_lock(&r8a66597->lock);
1222 		}
1223 
1224 		control_end(r8a66597, 1);
1225 
1226 		req = get_request_from_ep(ep);
1227 		if (ep->busy) {
1228 			ep->busy = 0;
1229 			if (list_empty(&ep->queue))
1230 				break;
1231 			start_packet(ep, req);
1232 		} else if (!list_empty(&ep->queue))
1233 			pipe_start(r8a66597, ep->pipenum);
1234 		}
1235 		break;
1236 	default:
1237 		pipe_stall(r8a66597, 0);
1238 		break;
1239 	}
1240 }
1241 
1242 static void set_feature(struct r8a66597 *r8a66597, struct usb_ctrlrequest *ctrl)
1243 {
1244 	u16 tmp;
1245 	int timeout = 3000;
1246 
1247 	switch (ctrl->bRequestType & USB_RECIP_MASK) {
1248 	case USB_RECIP_DEVICE:
1249 		switch (le16_to_cpu(ctrl->wValue)) {
1250 		case USB_DEVICE_TEST_MODE:
1251 			control_end(r8a66597, 1);
1252 			/* Wait for the completion of status stage */
1253 			do {
1254 				tmp = r8a66597_read(r8a66597, INTSTS0) & CTSQ;
1255 				udelay(1);
1256 			} while (tmp != CS_IDST || timeout-- > 0);
1257 
1258 			if (tmp == CS_IDST)
1259 				r8a66597_bset(r8a66597,
1260 					      le16_to_cpu(ctrl->wIndex >> 8),
1261 					      TESTMODE);
1262 			break;
1263 		default:
1264 			pipe_stall(r8a66597, 0);
1265 			break;
1266 		}
1267 		break;
1268 	case USB_RECIP_INTERFACE:
1269 		control_end(r8a66597, 1);
1270 		break;
1271 	case USB_RECIP_ENDPOINT: {
1272 		struct r8a66597_ep *ep;
1273 		u16 w_index = le16_to_cpu(ctrl->wIndex);
1274 
1275 		ep = r8a66597->epaddr2ep[w_index & USB_ENDPOINT_NUMBER_MASK];
1276 		pipe_stall(r8a66597, ep->pipenum);
1277 
1278 		control_end(r8a66597, 1);
1279 		}
1280 		break;
1281 	default:
1282 		pipe_stall(r8a66597, 0);
1283 		break;
1284 	}
1285 }
1286 
1287 /* if return value is true, call class driver's setup() */
1288 static int setup_packet(struct r8a66597 *r8a66597, struct usb_ctrlrequest *ctrl)
1289 {
1290 	u16 *p = (u16 *)ctrl;
1291 	unsigned long offset = USBREQ;
1292 	int i, ret = 0;
1293 
1294 	/* read fifo */
1295 	r8a66597_write(r8a66597, ~VALID, INTSTS0);
1296 
1297 	for (i = 0; i < 4; i++)
1298 		p[i] = r8a66597_read(r8a66597, offset + i*2);
1299 
1300 	/* check request */
1301 	if ((ctrl->bRequestType & USB_TYPE_MASK) == USB_TYPE_STANDARD) {
1302 		switch (ctrl->bRequest) {
1303 		case USB_REQ_GET_STATUS:
1304 			get_status(r8a66597, ctrl);
1305 			break;
1306 		case USB_REQ_CLEAR_FEATURE:
1307 			clear_feature(r8a66597, ctrl);
1308 			break;
1309 		case USB_REQ_SET_FEATURE:
1310 			set_feature(r8a66597, ctrl);
1311 			break;
1312 		default:
1313 			ret = 1;
1314 			break;
1315 		}
1316 	} else
1317 		ret = 1;
1318 	return ret;
1319 }
1320 
1321 static void r8a66597_update_usb_speed(struct r8a66597 *r8a66597)
1322 {
1323 	u16 speed = get_usb_speed(r8a66597);
1324 
1325 	switch (speed) {
1326 	case HSMODE:
1327 		r8a66597->gadget.speed = USB_SPEED_HIGH;
1328 		break;
1329 	case FSMODE:
1330 		r8a66597->gadget.speed = USB_SPEED_FULL;
1331 		break;
1332 	default:
1333 		r8a66597->gadget.speed = USB_SPEED_UNKNOWN;
1334 		dev_err(r8a66597_to_dev(r8a66597), "USB speed unknown\n");
1335 	}
1336 }
1337 
1338 static void irq_device_state(struct r8a66597 *r8a66597)
1339 {
1340 	u16 dvsq;
1341 
1342 	dvsq = r8a66597_read(r8a66597, INTSTS0) & DVSQ;
1343 	r8a66597_write(r8a66597, ~DVST, INTSTS0);
1344 
1345 	if (dvsq == DS_DFLT) {
1346 		/* bus reset */
1347 		spin_unlock(&r8a66597->lock);
1348 		usb_gadget_udc_reset(&r8a66597->gadget, r8a66597->driver);
1349 		spin_lock(&r8a66597->lock);
1350 		r8a66597_update_usb_speed(r8a66597);
1351 	}
1352 	if (r8a66597->old_dvsq == DS_CNFG && dvsq != DS_CNFG)
1353 		r8a66597_update_usb_speed(r8a66597);
1354 	if ((dvsq == DS_CNFG || dvsq == DS_ADDS)
1355 			&& r8a66597->gadget.speed == USB_SPEED_UNKNOWN)
1356 		r8a66597_update_usb_speed(r8a66597);
1357 
1358 	r8a66597->old_dvsq = dvsq;
1359 }
1360 
1361 static void irq_control_stage(struct r8a66597 *r8a66597)
1362 __releases(r8a66597->lock)
1363 __acquires(r8a66597->lock)
1364 {
1365 	struct usb_ctrlrequest ctrl;
1366 	u16 ctsq;
1367 
1368 	ctsq = r8a66597_read(r8a66597, INTSTS0) & CTSQ;
1369 	r8a66597_write(r8a66597, ~CTRT, INTSTS0);
1370 
1371 	switch (ctsq) {
1372 	case CS_IDST: {
1373 		struct r8a66597_ep *ep;
1374 		struct r8a66597_request *req;
1375 		ep = &r8a66597->ep[0];
1376 		req = get_request_from_ep(ep);
1377 		transfer_complete(ep, req, 0);
1378 		}
1379 		break;
1380 
1381 	case CS_RDDS:
1382 	case CS_WRDS:
1383 	case CS_WRND:
1384 		if (setup_packet(r8a66597, &ctrl)) {
1385 			spin_unlock(&r8a66597->lock);
1386 			if (r8a66597->driver->setup(&r8a66597->gadget, &ctrl)
1387 				< 0)
1388 				pipe_stall(r8a66597, 0);
1389 			spin_lock(&r8a66597->lock);
1390 		}
1391 		break;
1392 	case CS_RDSS:
1393 	case CS_WRSS:
1394 		control_end(r8a66597, 0);
1395 		break;
1396 	default:
1397 		dev_err(r8a66597_to_dev(r8a66597),
1398 			"ctrl_stage: unexpect ctsq(%x)\n", ctsq);
1399 		break;
1400 	}
1401 }
1402 
1403 static void sudmac_finish(struct r8a66597 *r8a66597, struct r8a66597_ep *ep)
1404 {
1405 	u16 pipenum;
1406 	struct r8a66597_request *req;
1407 	u32 len;
1408 	int i = 0;
1409 
1410 	pipenum = ep->pipenum;
1411 	pipe_change(r8a66597, pipenum);
1412 
1413 	while (!(r8a66597_read(r8a66597, ep->fifoctr) & FRDY)) {
1414 		udelay(1);
1415 		if (unlikely(i++ >= 10000)) {	/* timeout = 10 msec */
1416 			dev_err(r8a66597_to_dev(r8a66597),
1417 				"%s: FRDY was not set (%d)\n",
1418 				__func__, pipenum);
1419 			return;
1420 		}
1421 	}
1422 
1423 	r8a66597_bset(r8a66597, BCLR, ep->fifoctr);
1424 	req = get_request_from_ep(ep);
1425 
1426 	/* prepare parameters */
1427 	len = r8a66597_sudmac_read(r8a66597, CH0CBC);
1428 	req->req.actual += len;
1429 
1430 	/* clear */
1431 	r8a66597_sudmac_write(r8a66597, CH0STCLR, DSTSCLR);
1432 
1433 	/* check transfer finish */
1434 	if ((!req->req.zero && (req->req.actual == req->req.length))
1435 			|| (len % ep->ep.maxpacket)) {
1436 		if (ep->dma->dir) {
1437 			disable_irq_ready(r8a66597, pipenum);
1438 			enable_irq_empty(r8a66597, pipenum);
1439 		} else {
1440 			/* Clear the interrupt flag for next transfer */
1441 			r8a66597_write(r8a66597, ~(1 << pipenum), BRDYSTS);
1442 			transfer_complete(ep, req, 0);
1443 		}
1444 	}
1445 }
1446 
1447 static void r8a66597_sudmac_irq(struct r8a66597 *r8a66597)
1448 {
1449 	u32 irqsts;
1450 	struct r8a66597_ep *ep;
1451 	u16 pipenum;
1452 
1453 	irqsts = r8a66597_sudmac_read(r8a66597, DINTSTS);
1454 	if (irqsts & CH0ENDS) {
1455 		r8a66597_sudmac_write(r8a66597, CH0ENDC, DINTSTSCLR);
1456 		pipenum = (r8a66597_read(r8a66597, D0FIFOSEL) & CURPIPE);
1457 		ep = r8a66597->pipenum2ep[pipenum];
1458 		sudmac_finish(r8a66597, ep);
1459 	}
1460 }
1461 
1462 static irqreturn_t r8a66597_irq(int irq, void *_r8a66597)
1463 {
1464 	struct r8a66597 *r8a66597 = _r8a66597;
1465 	u16 intsts0;
1466 	u16 intenb0;
1467 	u16 savepipe;
1468 	u16 mask0;
1469 
1470 	spin_lock(&r8a66597->lock);
1471 
1472 	if (r8a66597_is_sudmac(r8a66597))
1473 		r8a66597_sudmac_irq(r8a66597);
1474 
1475 	intsts0 = r8a66597_read(r8a66597, INTSTS0);
1476 	intenb0 = r8a66597_read(r8a66597, INTENB0);
1477 
1478 	savepipe = r8a66597_read(r8a66597, CFIFOSEL);
1479 
1480 	mask0 = intsts0 & intenb0;
1481 	if (mask0) {
1482 		u16 brdysts = r8a66597_read(r8a66597, BRDYSTS);
1483 		u16 bempsts = r8a66597_read(r8a66597, BEMPSTS);
1484 		u16 brdyenb = r8a66597_read(r8a66597, BRDYENB);
1485 		u16 bempenb = r8a66597_read(r8a66597, BEMPENB);
1486 
1487 		if (mask0 & VBINT) {
1488 			r8a66597_write(r8a66597,  0xffff & ~VBINT,
1489 					INTSTS0);
1490 			r8a66597_start_xclock(r8a66597);
1491 
1492 			/* start vbus sampling */
1493 			r8a66597->old_vbus = r8a66597_read(r8a66597, INTSTS0)
1494 					& VBSTS;
1495 			r8a66597->scount = R8A66597_MAX_SAMPLING;
1496 
1497 			mod_timer(&r8a66597->timer,
1498 					jiffies + msecs_to_jiffies(50));
1499 		}
1500 		if (intsts0 & DVSQ)
1501 			irq_device_state(r8a66597);
1502 
1503 		if ((intsts0 & BRDY) && (intenb0 & BRDYE)
1504 				&& (brdysts & brdyenb))
1505 			irq_pipe_ready(r8a66597, brdysts, brdyenb);
1506 		if ((intsts0 & BEMP) && (intenb0 & BEMPE)
1507 				&& (bempsts & bempenb))
1508 			irq_pipe_empty(r8a66597, bempsts, bempenb);
1509 
1510 		if (intsts0 & CTRT)
1511 			irq_control_stage(r8a66597);
1512 	}
1513 
1514 	r8a66597_write(r8a66597, savepipe, CFIFOSEL);
1515 
1516 	spin_unlock(&r8a66597->lock);
1517 	return IRQ_HANDLED;
1518 }
1519 
1520 static void r8a66597_timer(unsigned long _r8a66597)
1521 {
1522 	struct r8a66597 *r8a66597 = (struct r8a66597 *)_r8a66597;
1523 	unsigned long flags;
1524 	u16 tmp;
1525 
1526 	spin_lock_irqsave(&r8a66597->lock, flags);
1527 	tmp = r8a66597_read(r8a66597, SYSCFG0);
1528 	if (r8a66597->scount > 0) {
1529 		tmp = r8a66597_read(r8a66597, INTSTS0) & VBSTS;
1530 		if (tmp == r8a66597->old_vbus) {
1531 			r8a66597->scount--;
1532 			if (r8a66597->scount == 0) {
1533 				if (tmp == VBSTS)
1534 					r8a66597_usb_connect(r8a66597);
1535 				else
1536 					r8a66597_usb_disconnect(r8a66597);
1537 			} else {
1538 				mod_timer(&r8a66597->timer,
1539 					jiffies + msecs_to_jiffies(50));
1540 			}
1541 		} else {
1542 			r8a66597->scount = R8A66597_MAX_SAMPLING;
1543 			r8a66597->old_vbus = tmp;
1544 			mod_timer(&r8a66597->timer,
1545 					jiffies + msecs_to_jiffies(50));
1546 		}
1547 	}
1548 	spin_unlock_irqrestore(&r8a66597->lock, flags);
1549 }
1550 
1551 /*-------------------------------------------------------------------------*/
1552 static int r8a66597_enable(struct usb_ep *_ep,
1553 			 const struct usb_endpoint_descriptor *desc)
1554 {
1555 	struct r8a66597_ep *ep;
1556 
1557 	ep = container_of(_ep, struct r8a66597_ep, ep);
1558 	return alloc_pipe_config(ep, desc);
1559 }
1560 
1561 static int r8a66597_disable(struct usb_ep *_ep)
1562 {
1563 	struct r8a66597_ep *ep;
1564 	struct r8a66597_request *req;
1565 	unsigned long flags;
1566 
1567 	ep = container_of(_ep, struct r8a66597_ep, ep);
1568 	BUG_ON(!ep);
1569 
1570 	while (!list_empty(&ep->queue)) {
1571 		req = get_request_from_ep(ep);
1572 		spin_lock_irqsave(&ep->r8a66597->lock, flags);
1573 		transfer_complete(ep, req, -ECONNRESET);
1574 		spin_unlock_irqrestore(&ep->r8a66597->lock, flags);
1575 	}
1576 
1577 	pipe_irq_disable(ep->r8a66597, ep->pipenum);
1578 	return free_pipe_config(ep);
1579 }
1580 
1581 static struct usb_request *r8a66597_alloc_request(struct usb_ep *_ep,
1582 						gfp_t gfp_flags)
1583 {
1584 	struct r8a66597_request *req;
1585 
1586 	req = kzalloc(sizeof(struct r8a66597_request), gfp_flags);
1587 	if (!req)
1588 		return NULL;
1589 
1590 	INIT_LIST_HEAD(&req->queue);
1591 
1592 	return &req->req;
1593 }
1594 
1595 static void r8a66597_free_request(struct usb_ep *_ep, struct usb_request *_req)
1596 {
1597 	struct r8a66597_request *req;
1598 
1599 	req = container_of(_req, struct r8a66597_request, req);
1600 	kfree(req);
1601 }
1602 
1603 static int r8a66597_queue(struct usb_ep *_ep, struct usb_request *_req,
1604 			gfp_t gfp_flags)
1605 {
1606 	struct r8a66597_ep *ep;
1607 	struct r8a66597_request *req;
1608 	unsigned long flags;
1609 	int request = 0;
1610 
1611 	ep = container_of(_ep, struct r8a66597_ep, ep);
1612 	req = container_of(_req, struct r8a66597_request, req);
1613 
1614 	if (ep->r8a66597->gadget.speed == USB_SPEED_UNKNOWN)
1615 		return -ESHUTDOWN;
1616 
1617 	spin_lock_irqsave(&ep->r8a66597->lock, flags);
1618 
1619 	if (list_empty(&ep->queue))
1620 		request = 1;
1621 
1622 	list_add_tail(&req->queue, &ep->queue);
1623 	req->req.actual = 0;
1624 	req->req.status = -EINPROGRESS;
1625 
1626 	if (ep->ep.desc == NULL)	/* control */
1627 		start_ep0(ep, req);
1628 	else {
1629 		if (request && !ep->busy)
1630 			start_packet(ep, req);
1631 	}
1632 
1633 	spin_unlock_irqrestore(&ep->r8a66597->lock, flags);
1634 
1635 	return 0;
1636 }
1637 
1638 static int r8a66597_dequeue(struct usb_ep *_ep, struct usb_request *_req)
1639 {
1640 	struct r8a66597_ep *ep;
1641 	struct r8a66597_request *req;
1642 	unsigned long flags;
1643 
1644 	ep = container_of(_ep, struct r8a66597_ep, ep);
1645 	req = container_of(_req, struct r8a66597_request, req);
1646 
1647 	spin_lock_irqsave(&ep->r8a66597->lock, flags);
1648 	if (!list_empty(&ep->queue))
1649 		transfer_complete(ep, req, -ECONNRESET);
1650 	spin_unlock_irqrestore(&ep->r8a66597->lock, flags);
1651 
1652 	return 0;
1653 }
1654 
1655 static int r8a66597_set_halt(struct usb_ep *_ep, int value)
1656 {
1657 	struct r8a66597_ep *ep = container_of(_ep, struct r8a66597_ep, ep);
1658 	unsigned long flags;
1659 	int ret = 0;
1660 
1661 	spin_lock_irqsave(&ep->r8a66597->lock, flags);
1662 	if (!list_empty(&ep->queue)) {
1663 		ret = -EAGAIN;
1664 	} else if (value) {
1665 		ep->busy = 1;
1666 		pipe_stall(ep->r8a66597, ep->pipenum);
1667 	} else {
1668 		ep->busy = 0;
1669 		ep->wedge = 0;
1670 		pipe_stop(ep->r8a66597, ep->pipenum);
1671 	}
1672 	spin_unlock_irqrestore(&ep->r8a66597->lock, flags);
1673 	return ret;
1674 }
1675 
1676 static int r8a66597_set_wedge(struct usb_ep *_ep)
1677 {
1678 	struct r8a66597_ep *ep;
1679 	unsigned long flags;
1680 
1681 	ep = container_of(_ep, struct r8a66597_ep, ep);
1682 
1683 	if (!ep || !ep->ep.desc)
1684 		return -EINVAL;
1685 
1686 	spin_lock_irqsave(&ep->r8a66597->lock, flags);
1687 	ep->wedge = 1;
1688 	spin_unlock_irqrestore(&ep->r8a66597->lock, flags);
1689 
1690 	return usb_ep_set_halt(_ep);
1691 }
1692 
1693 static void r8a66597_fifo_flush(struct usb_ep *_ep)
1694 {
1695 	struct r8a66597_ep *ep;
1696 	unsigned long flags;
1697 
1698 	ep = container_of(_ep, struct r8a66597_ep, ep);
1699 	spin_lock_irqsave(&ep->r8a66597->lock, flags);
1700 	if (list_empty(&ep->queue) && !ep->busy) {
1701 		pipe_stop(ep->r8a66597, ep->pipenum);
1702 		r8a66597_bclr(ep->r8a66597, BCLR, ep->fifoctr);
1703 		r8a66597_write(ep->r8a66597, ACLRM, ep->pipectr);
1704 		r8a66597_write(ep->r8a66597, 0, ep->pipectr);
1705 	}
1706 	spin_unlock_irqrestore(&ep->r8a66597->lock, flags);
1707 }
1708 
1709 static struct usb_ep_ops r8a66597_ep_ops = {
1710 	.enable		= r8a66597_enable,
1711 	.disable	= r8a66597_disable,
1712 
1713 	.alloc_request	= r8a66597_alloc_request,
1714 	.free_request	= r8a66597_free_request,
1715 
1716 	.queue		= r8a66597_queue,
1717 	.dequeue	= r8a66597_dequeue,
1718 
1719 	.set_halt	= r8a66597_set_halt,
1720 	.set_wedge	= r8a66597_set_wedge,
1721 	.fifo_flush	= r8a66597_fifo_flush,
1722 };
1723 
1724 /*-------------------------------------------------------------------------*/
1725 static int r8a66597_start(struct usb_gadget *gadget,
1726 		struct usb_gadget_driver *driver)
1727 {
1728 	struct r8a66597 *r8a66597 = gadget_to_r8a66597(gadget);
1729 
1730 	if (!driver
1731 			|| driver->max_speed < USB_SPEED_HIGH
1732 			|| !driver->setup)
1733 		return -EINVAL;
1734 	if (!r8a66597)
1735 		return -ENODEV;
1736 
1737 	/* hook up the driver */
1738 	r8a66597->driver = driver;
1739 
1740 	init_controller(r8a66597);
1741 	r8a66597_bset(r8a66597, VBSE, INTENB0);
1742 	if (r8a66597_read(r8a66597, INTSTS0) & VBSTS) {
1743 		r8a66597_start_xclock(r8a66597);
1744 		/* start vbus sampling */
1745 		r8a66597->old_vbus = r8a66597_read(r8a66597,
1746 					 INTSTS0) & VBSTS;
1747 		r8a66597->scount = R8A66597_MAX_SAMPLING;
1748 		mod_timer(&r8a66597->timer, jiffies + msecs_to_jiffies(50));
1749 	}
1750 
1751 	return 0;
1752 }
1753 
1754 static int r8a66597_stop(struct usb_gadget *gadget)
1755 {
1756 	struct r8a66597 *r8a66597 = gadget_to_r8a66597(gadget);
1757 	unsigned long flags;
1758 
1759 	spin_lock_irqsave(&r8a66597->lock, flags);
1760 	r8a66597_bclr(r8a66597, VBSE, INTENB0);
1761 	disable_controller(r8a66597);
1762 	spin_unlock_irqrestore(&r8a66597->lock, flags);
1763 
1764 	r8a66597->driver = NULL;
1765 	return 0;
1766 }
1767 
1768 /*-------------------------------------------------------------------------*/
1769 static int r8a66597_get_frame(struct usb_gadget *_gadget)
1770 {
1771 	struct r8a66597 *r8a66597 = gadget_to_r8a66597(_gadget);
1772 	return r8a66597_read(r8a66597, FRMNUM) & 0x03FF;
1773 }
1774 
1775 static int r8a66597_pullup(struct usb_gadget *gadget, int is_on)
1776 {
1777 	struct r8a66597 *r8a66597 = gadget_to_r8a66597(gadget);
1778 	unsigned long flags;
1779 
1780 	spin_lock_irqsave(&r8a66597->lock, flags);
1781 	if (is_on)
1782 		r8a66597_bset(r8a66597, DPRPU, SYSCFG0);
1783 	else
1784 		r8a66597_bclr(r8a66597, DPRPU, SYSCFG0);
1785 	spin_unlock_irqrestore(&r8a66597->lock, flags);
1786 
1787 	return 0;
1788 }
1789 
1790 static int r8a66597_set_selfpowered(struct usb_gadget *gadget, int is_self)
1791 {
1792 	struct r8a66597 *r8a66597 = gadget_to_r8a66597(gadget);
1793 
1794 	gadget->is_selfpowered = (is_self != 0);
1795 	if (is_self)
1796 		r8a66597->device_status |= 1 << USB_DEVICE_SELF_POWERED;
1797 	else
1798 		r8a66597->device_status &= ~(1 << USB_DEVICE_SELF_POWERED);
1799 
1800 	return 0;
1801 }
1802 
1803 static const struct usb_gadget_ops r8a66597_gadget_ops = {
1804 	.get_frame		= r8a66597_get_frame,
1805 	.udc_start		= r8a66597_start,
1806 	.udc_stop		= r8a66597_stop,
1807 	.pullup			= r8a66597_pullup,
1808 	.set_selfpowered	= r8a66597_set_selfpowered,
1809 };
1810 
1811 static int r8a66597_remove(struct platform_device *pdev)
1812 {
1813 	struct r8a66597		*r8a66597 = platform_get_drvdata(pdev);
1814 
1815 	usb_del_gadget_udc(&r8a66597->gadget);
1816 	del_timer_sync(&r8a66597->timer);
1817 	r8a66597_free_request(&r8a66597->ep[0].ep, r8a66597->ep0_req);
1818 
1819 	if (r8a66597->pdata->on_chip) {
1820 		clk_disable_unprepare(r8a66597->clk);
1821 	}
1822 
1823 	return 0;
1824 }
1825 
1826 static void nop_completion(struct usb_ep *ep, struct usb_request *r)
1827 {
1828 }
1829 
1830 static int r8a66597_sudmac_ioremap(struct r8a66597 *r8a66597,
1831 					  struct platform_device *pdev)
1832 {
1833 	struct resource *res;
1834 
1835 	res = platform_get_resource_byname(pdev, IORESOURCE_MEM, "sudmac");
1836 	r8a66597->sudmac_reg = devm_ioremap_resource(&pdev->dev, res);
1837 	return PTR_ERR_OR_ZERO(r8a66597->sudmac_reg);
1838 }
1839 
1840 static int r8a66597_probe(struct platform_device *pdev)
1841 {
1842 	struct device *dev = &pdev->dev;
1843 	char clk_name[8];
1844 	struct resource *res, *ires;
1845 	int irq;
1846 	void __iomem *reg = NULL;
1847 	struct r8a66597 *r8a66597 = NULL;
1848 	int ret = 0;
1849 	int i;
1850 	unsigned long irq_trigger;
1851 
1852 	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
1853 	reg = devm_ioremap_resource(&pdev->dev, res);
1854 	if (IS_ERR(reg))
1855 		return PTR_ERR(reg);
1856 
1857 	ires = platform_get_resource(pdev, IORESOURCE_IRQ, 0);
1858 	irq = ires->start;
1859 	irq_trigger = ires->flags & IRQF_TRIGGER_MASK;
1860 
1861 	if (irq < 0) {
1862 		dev_err(dev, "platform_get_irq error.\n");
1863 		return -ENODEV;
1864 	}
1865 
1866 	/* initialize ucd */
1867 	r8a66597 = devm_kzalloc(dev, sizeof(struct r8a66597), GFP_KERNEL);
1868 	if (r8a66597 == NULL)
1869 		return -ENOMEM;
1870 
1871 	spin_lock_init(&r8a66597->lock);
1872 	platform_set_drvdata(pdev, r8a66597);
1873 	r8a66597->pdata = dev_get_platdata(dev);
1874 	r8a66597->irq_sense_low = irq_trigger == IRQF_TRIGGER_LOW;
1875 
1876 	r8a66597->gadget.ops = &r8a66597_gadget_ops;
1877 	r8a66597->gadget.max_speed = USB_SPEED_HIGH;
1878 	r8a66597->gadget.name = udc_name;
1879 
1880 	init_timer(&r8a66597->timer);
1881 	r8a66597->timer.function = r8a66597_timer;
1882 	r8a66597->timer.data = (unsigned long)r8a66597;
1883 	r8a66597->reg = reg;
1884 
1885 	if (r8a66597->pdata->on_chip) {
1886 		snprintf(clk_name, sizeof(clk_name), "usb%d", pdev->id);
1887 		r8a66597->clk = devm_clk_get(dev, clk_name);
1888 		if (IS_ERR(r8a66597->clk)) {
1889 			dev_err(dev, "cannot get clock \"%s\"\n", clk_name);
1890 			return PTR_ERR(r8a66597->clk);
1891 		}
1892 		clk_prepare_enable(r8a66597->clk);
1893 	}
1894 
1895 	if (r8a66597->pdata->sudmac) {
1896 		ret = r8a66597_sudmac_ioremap(r8a66597, pdev);
1897 		if (ret < 0)
1898 			goto clean_up2;
1899 	}
1900 
1901 	disable_controller(r8a66597); /* make sure controller is disabled */
1902 
1903 	ret = devm_request_irq(dev, irq, r8a66597_irq, IRQF_SHARED,
1904 			       udc_name, r8a66597);
1905 	if (ret < 0) {
1906 		dev_err(dev, "request_irq error (%d)\n", ret);
1907 		goto clean_up2;
1908 	}
1909 
1910 	INIT_LIST_HEAD(&r8a66597->gadget.ep_list);
1911 	r8a66597->gadget.ep0 = &r8a66597->ep[0].ep;
1912 	INIT_LIST_HEAD(&r8a66597->gadget.ep0->ep_list);
1913 	for (i = 0; i < R8A66597_MAX_NUM_PIPE; i++) {
1914 		struct r8a66597_ep *ep = &r8a66597->ep[i];
1915 
1916 		if (i != 0) {
1917 			INIT_LIST_HEAD(&r8a66597->ep[i].ep.ep_list);
1918 			list_add_tail(&r8a66597->ep[i].ep.ep_list,
1919 					&r8a66597->gadget.ep_list);
1920 		}
1921 		ep->r8a66597 = r8a66597;
1922 		INIT_LIST_HEAD(&ep->queue);
1923 		ep->ep.name = r8a66597_ep_name[i];
1924 		ep->ep.ops = &r8a66597_ep_ops;
1925 		usb_ep_set_maxpacket_limit(&ep->ep, 512);
1926 
1927 		if (i == 0) {
1928 			ep->ep.caps.type_control = true;
1929 		} else {
1930 			ep->ep.caps.type_iso = true;
1931 			ep->ep.caps.type_bulk = true;
1932 			ep->ep.caps.type_int = true;
1933 		}
1934 		ep->ep.caps.dir_in = true;
1935 		ep->ep.caps.dir_out = true;
1936 	}
1937 	usb_ep_set_maxpacket_limit(&r8a66597->ep[0].ep, 64);
1938 	r8a66597->ep[0].pipenum = 0;
1939 	r8a66597->ep[0].fifoaddr = CFIFO;
1940 	r8a66597->ep[0].fifosel = CFIFOSEL;
1941 	r8a66597->ep[0].fifoctr = CFIFOCTR;
1942 	r8a66597->ep[0].pipectr = get_pipectr_addr(0);
1943 	r8a66597->pipenum2ep[0] = &r8a66597->ep[0];
1944 	r8a66597->epaddr2ep[0] = &r8a66597->ep[0];
1945 
1946 	r8a66597->ep0_req = r8a66597_alloc_request(&r8a66597->ep[0].ep,
1947 							GFP_KERNEL);
1948 	if (r8a66597->ep0_req == NULL) {
1949 		ret = -ENOMEM;
1950 		goto clean_up2;
1951 	}
1952 	r8a66597->ep0_req->complete = nop_completion;
1953 
1954 	ret = usb_add_gadget_udc(dev, &r8a66597->gadget);
1955 	if (ret)
1956 		goto err_add_udc;
1957 
1958 	dev_info(dev, "version %s\n", DRIVER_VERSION);
1959 	return 0;
1960 
1961 err_add_udc:
1962 	r8a66597_free_request(&r8a66597->ep[0].ep, r8a66597->ep0_req);
1963 clean_up2:
1964 	if (r8a66597->pdata->on_chip)
1965 		clk_disable_unprepare(r8a66597->clk);
1966 
1967 	if (r8a66597->ep0_req)
1968 		r8a66597_free_request(&r8a66597->ep[0].ep, r8a66597->ep0_req);
1969 
1970 	return ret;
1971 }
1972 
1973 /*-------------------------------------------------------------------------*/
1974 static struct platform_driver r8a66597_driver = {
1975 	.remove =	r8a66597_remove,
1976 	.driver		= {
1977 		.name =	(char *) udc_name,
1978 	},
1979 };
1980 
1981 module_platform_driver_probe(r8a66597_driver, r8a66597_probe);
1982 
1983 MODULE_DESCRIPTION("R8A66597 USB gadget driver");
1984 MODULE_LICENSE("GPL");
1985 MODULE_AUTHOR("Yoshihiro Shimoda");
1986 MODULE_ALIAS("platform:r8a66597_udc");
1987