xref: /openbmc/linux/net/9p/trans_fd.c (revision 2eb5f31b)
1 /*
2  * linux/fs/9p/trans_fd.c
3  *
4  * Fd transport layer.  Includes deprecated socket layer.
5  *
6  *  Copyright (C) 2006 by Russ Cox <rsc@swtch.com>
7  *  Copyright (C) 2004-2005 by Latchesar Ionkov <lucho@ionkov.net>
8  *  Copyright (C) 2004-2008 by Eric Van Hensbergen <ericvh@gmail.com>
9  *  Copyright (C) 1997-2002 by Ron Minnich <rminnich@sarnoff.com>
10  *
11  *  This program is free software; you can redistribute it and/or modify
12  *  it under the terms of the GNU General Public License version 2
13  *  as published by the Free Software Foundation.
14  *
15  *  This program is distributed in the hope that it will be useful,
16  *  but WITHOUT ANY WARRANTY; without even the implied warranty of
17  *  MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
18  *  GNU General Public License for more details.
19  *
20  *  You should have received a copy of the GNU General Public License
21  *  along with this program; if not, write to:
22  *  Free Software Foundation
23  *  51 Franklin Street, Fifth Floor
24  *  Boston, MA  02111-1301  USA
25  *
26  */
27 
28 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
29 
30 #include <linux/in.h>
31 #include <linux/module.h>
32 #include <linux/net.h>
33 #include <linux/ipv6.h>
34 #include <linux/kthread.h>
35 #include <linux/errno.h>
36 #include <linux/kernel.h>
37 #include <linux/un.h>
38 #include <linux/uaccess.h>
39 #include <linux/inet.h>
40 #include <linux/idr.h>
41 #include <linux/file.h>
42 #include <linux/parser.h>
43 #include <linux/slab.h>
44 #include <net/9p/9p.h>
45 #include <net/9p/client.h>
46 #include <net/9p/transport.h>
47 
48 #include <linux/syscalls.h> /* killme */
49 
50 #define P9_PORT 564
51 #define MAX_SOCK_BUF (64*1024)
52 #define MAXPOLLWADDR	2
53 
54 /**
55  * struct p9_fd_opts - per-transport options
56  * @rfd: file descriptor for reading (trans=fd)
57  * @wfd: file descriptor for writing (trans=fd)
58  * @port: port to connect to (trans=tcp)
59  *
60  */
61 
62 struct p9_fd_opts {
63 	int rfd;
64 	int wfd;
65 	u16 port;
66 	int privport;
67 };
68 
69 /*
70   * Option Parsing (code inspired by NFS code)
71   *  - a little lazy - parse all fd-transport options
72   */
73 
74 enum {
75 	/* Options that take integer arguments */
76 	Opt_port, Opt_rfdno, Opt_wfdno, Opt_err,
77 	/* Options that take no arguments */
78 	Opt_privport,
79 };
80 
81 static const match_table_t tokens = {
82 	{Opt_port, "port=%u"},
83 	{Opt_rfdno, "rfdno=%u"},
84 	{Opt_wfdno, "wfdno=%u"},
85 	{Opt_privport, "privport"},
86 	{Opt_err, NULL},
87 };
88 
89 enum {
90 	Rworksched = 1,		/* read work scheduled or running */
91 	Rpending = 2,		/* can read */
92 	Wworksched = 4,		/* write work scheduled or running */
93 	Wpending = 8,		/* can write */
94 };
95 
96 struct p9_poll_wait {
97 	struct p9_conn *conn;
98 	wait_queue_t wait;
99 	wait_queue_head_t *wait_addr;
100 };
101 
102 /**
103  * struct p9_conn - fd mux connection state information
104  * @mux_list: list link for mux to manage multiple connections (?)
105  * @client: reference to client instance for this connection
106  * @err: error state
107  * @req_list: accounting for requests which have been sent
108  * @unsent_req_list: accounting for requests that haven't been sent
109  * @req: current request being processed (if any)
110  * @tmp_buf: temporary buffer to read in header
111  * @rsize: amount to read for current frame
112  * @rpos: read position in current frame
113  * @rbuf: current read buffer
114  * @wpos: write position for current frame
115  * @wsize: amount of data to write for current frame
116  * @wbuf: current write buffer
117  * @poll_pending_link: pending links to be polled per conn
118  * @poll_wait: array of wait_q's for various worker threads
119  * @pt: poll state
120  * @rq: current read work
121  * @wq: current write work
122  * @wsched: ????
123  *
124  */
125 
126 struct p9_conn {
127 	struct list_head mux_list;
128 	struct p9_client *client;
129 	int err;
130 	struct list_head req_list;
131 	struct list_head unsent_req_list;
132 	struct p9_req_t *req;
133 	char tmp_buf[7];
134 	int rsize;
135 	int rpos;
136 	char *rbuf;
137 	int wpos;
138 	int wsize;
139 	char *wbuf;
140 	struct list_head poll_pending_link;
141 	struct p9_poll_wait poll_wait[MAXPOLLWADDR];
142 	poll_table pt;
143 	struct work_struct rq;
144 	struct work_struct wq;
145 	unsigned long wsched;
146 };
147 
148 /**
149  * struct p9_trans_fd - transport state
150  * @rd: reference to file to read from
151  * @wr: reference of file to write to
152  * @conn: connection state reference
153  *
154  */
155 
156 struct p9_trans_fd {
157 	struct file *rd;
158 	struct file *wr;
159 	struct p9_conn conn;
160 };
161 
162 static void p9_poll_workfn(struct work_struct *work);
163 
164 static DEFINE_SPINLOCK(p9_poll_lock);
165 static LIST_HEAD(p9_poll_pending_list);
166 static DECLARE_WORK(p9_poll_work, p9_poll_workfn);
167 
168 static unsigned int p9_ipport_resv_min = P9_DEF_MIN_RESVPORT;
169 static unsigned int p9_ipport_resv_max = P9_DEF_MAX_RESVPORT;
170 
171 static void p9_mux_poll_stop(struct p9_conn *m)
172 {
173 	unsigned long flags;
174 	int i;
175 
176 	for (i = 0; i < ARRAY_SIZE(m->poll_wait); i++) {
177 		struct p9_poll_wait *pwait = &m->poll_wait[i];
178 
179 		if (pwait->wait_addr) {
180 			remove_wait_queue(pwait->wait_addr, &pwait->wait);
181 			pwait->wait_addr = NULL;
182 		}
183 	}
184 
185 	spin_lock_irqsave(&p9_poll_lock, flags);
186 	list_del_init(&m->poll_pending_link);
187 	spin_unlock_irqrestore(&p9_poll_lock, flags);
188 }
189 
190 /**
191  * p9_conn_cancel - cancel all pending requests with error
192  * @m: mux data
193  * @err: error code
194  *
195  */
196 
197 static void p9_conn_cancel(struct p9_conn *m, int err)
198 {
199 	struct p9_req_t *req, *rtmp;
200 	unsigned long flags;
201 	LIST_HEAD(cancel_list);
202 
203 	p9_debug(P9_DEBUG_ERROR, "mux %p err %d\n", m, err);
204 
205 	spin_lock_irqsave(&m->client->lock, flags);
206 
207 	if (m->err) {
208 		spin_unlock_irqrestore(&m->client->lock, flags);
209 		return;
210 	}
211 
212 	m->err = err;
213 
214 	list_for_each_entry_safe(req, rtmp, &m->req_list, req_list) {
215 		list_move(&req->req_list, &cancel_list);
216 	}
217 	list_for_each_entry_safe(req, rtmp, &m->unsent_req_list, req_list) {
218 		list_move(&req->req_list, &cancel_list);
219 	}
220 	spin_unlock_irqrestore(&m->client->lock, flags);
221 
222 	list_for_each_entry_safe(req, rtmp, &cancel_list, req_list) {
223 		p9_debug(P9_DEBUG_ERROR, "call back req %p\n", req);
224 		list_del(&req->req_list);
225 		if (!req->t_err)
226 			req->t_err = err;
227 		p9_client_cb(m->client, req, REQ_STATUS_ERROR);
228 	}
229 }
230 
231 static int
232 p9_fd_poll(struct p9_client *client, struct poll_table_struct *pt)
233 {
234 	int ret, n;
235 	struct p9_trans_fd *ts = NULL;
236 
237 	if (client && client->status == Connected)
238 		ts = client->trans;
239 
240 	if (!ts)
241 		return -EREMOTEIO;
242 
243 	if (!ts->rd->f_op->poll)
244 		return -EIO;
245 
246 	if (!ts->wr->f_op->poll)
247 		return -EIO;
248 
249 	ret = ts->rd->f_op->poll(ts->rd, pt);
250 	if (ret < 0)
251 		return ret;
252 
253 	if (ts->rd != ts->wr) {
254 		n = ts->wr->f_op->poll(ts->wr, pt);
255 		if (n < 0)
256 			return n;
257 		ret = (ret & ~POLLOUT) | (n & ~POLLIN);
258 	}
259 
260 	return ret;
261 }
262 
263 /**
264  * p9_fd_read- read from a fd
265  * @client: client instance
266  * @v: buffer to receive data into
267  * @len: size of receive buffer
268  *
269  */
270 
271 static int p9_fd_read(struct p9_client *client, void *v, int len)
272 {
273 	int ret;
274 	struct p9_trans_fd *ts = NULL;
275 
276 	if (client && client->status != Disconnected)
277 		ts = client->trans;
278 
279 	if (!ts)
280 		return -EREMOTEIO;
281 
282 	if (!(ts->rd->f_flags & O_NONBLOCK))
283 		p9_debug(P9_DEBUG_ERROR, "blocking read ...\n");
284 
285 	ret = kernel_read(ts->rd, ts->rd->f_pos, v, len);
286 	if (ret <= 0 && ret != -ERESTARTSYS && ret != -EAGAIN)
287 		client->status = Disconnected;
288 	return ret;
289 }
290 
291 /**
292  * p9_read_work - called when there is some data to be read from a transport
293  * @work: container of work to be done
294  *
295  */
296 
297 static void p9_read_work(struct work_struct *work)
298 {
299 	int n, err;
300 	struct p9_conn *m;
301 	int status = REQ_STATUS_ERROR;
302 
303 	m = container_of(work, struct p9_conn, rq);
304 
305 	if (m->err < 0)
306 		return;
307 
308 	p9_debug(P9_DEBUG_TRANS, "start mux %p pos %d\n", m, m->rpos);
309 
310 	if (!m->rbuf) {
311 		m->rbuf = m->tmp_buf;
312 		m->rpos = 0;
313 		m->rsize = 7; /* start by reading header */
314 	}
315 
316 	clear_bit(Rpending, &m->wsched);
317 	p9_debug(P9_DEBUG_TRANS, "read mux %p pos %d size: %d = %d\n",
318 		 m, m->rpos, m->rsize, m->rsize-m->rpos);
319 	err = p9_fd_read(m->client, m->rbuf + m->rpos,
320 						m->rsize - m->rpos);
321 	p9_debug(P9_DEBUG_TRANS, "mux %p got %d bytes\n", m, err);
322 	if (err == -EAGAIN) {
323 		goto end_clear;
324 	}
325 
326 	if (err <= 0)
327 		goto error;
328 
329 	m->rpos += err;
330 
331 	if ((!m->req) && (m->rpos == m->rsize)) { /* header read in */
332 		u16 tag;
333 		p9_debug(P9_DEBUG_TRANS, "got new header\n");
334 
335 		n = le32_to_cpu(*(__le32 *) m->rbuf); /* read packet size */
336 		if (n >= m->client->msize) {
337 			p9_debug(P9_DEBUG_ERROR,
338 				 "requested packet size too big: %d\n", n);
339 			err = -EIO;
340 			goto error;
341 		}
342 
343 		tag = le16_to_cpu(*(__le16 *) (m->rbuf+5)); /* read tag */
344 		p9_debug(P9_DEBUG_TRANS,
345 			 "mux %p pkt: size: %d bytes tag: %d\n", m, n, tag);
346 
347 		m->req = p9_tag_lookup(m->client, tag);
348 		if (!m->req || (m->req->status != REQ_STATUS_SENT)) {
349 			p9_debug(P9_DEBUG_ERROR, "Unexpected packet tag %d\n",
350 				 tag);
351 			err = -EIO;
352 			goto error;
353 		}
354 
355 		if (m->req->rc == NULL) {
356 			m->req->rc = kmalloc(sizeof(struct p9_fcall) +
357 						m->client->msize, GFP_NOFS);
358 			if (!m->req->rc) {
359 				m->req = NULL;
360 				err = -ENOMEM;
361 				goto error;
362 			}
363 		}
364 		m->rbuf = (char *)m->req->rc + sizeof(struct p9_fcall);
365 		memcpy(m->rbuf, m->tmp_buf, m->rsize);
366 		m->rsize = n;
367 	}
368 
369 	/* not an else because some packets (like clunk) have no payload */
370 	if ((m->req) && (m->rpos == m->rsize)) { /* packet is read in */
371 		p9_debug(P9_DEBUG_TRANS, "got new packet\n");
372 		spin_lock(&m->client->lock);
373 		if (m->req->status != REQ_STATUS_ERROR)
374 			status = REQ_STATUS_RCVD;
375 		list_del(&m->req->req_list);
376 		spin_unlock(&m->client->lock);
377 		p9_client_cb(m->client, m->req, status);
378 		m->rbuf = NULL;
379 		m->rpos = 0;
380 		m->rsize = 0;
381 		m->req = NULL;
382 	}
383 
384 end_clear:
385 	clear_bit(Rworksched, &m->wsched);
386 
387 	if (!list_empty(&m->req_list)) {
388 		if (test_and_clear_bit(Rpending, &m->wsched))
389 			n = POLLIN;
390 		else
391 			n = p9_fd_poll(m->client, NULL);
392 
393 		if ((n & POLLIN) && !test_and_set_bit(Rworksched, &m->wsched)) {
394 			p9_debug(P9_DEBUG_TRANS, "sched read work %p\n", m);
395 			schedule_work(&m->rq);
396 		}
397 	}
398 
399 	return;
400 error:
401 	p9_conn_cancel(m, err);
402 	clear_bit(Rworksched, &m->wsched);
403 }
404 
405 /**
406  * p9_fd_write - write to a socket
407  * @client: client instance
408  * @v: buffer to send data from
409  * @len: size of send buffer
410  *
411  */
412 
413 static int p9_fd_write(struct p9_client *client, void *v, int len)
414 {
415 	int ret;
416 	mm_segment_t oldfs;
417 	struct p9_trans_fd *ts = NULL;
418 
419 	if (client && client->status != Disconnected)
420 		ts = client->trans;
421 
422 	if (!ts)
423 		return -EREMOTEIO;
424 
425 	if (!(ts->wr->f_flags & O_NONBLOCK))
426 		p9_debug(P9_DEBUG_ERROR, "blocking write ...\n");
427 
428 	oldfs = get_fs();
429 	set_fs(get_ds());
430 	/* The cast to a user pointer is valid due to the set_fs() */
431 	ret = vfs_write(ts->wr, (__force void __user *)v, len, &ts->wr->f_pos);
432 	set_fs(oldfs);
433 
434 	if (ret <= 0 && ret != -ERESTARTSYS && ret != -EAGAIN)
435 		client->status = Disconnected;
436 	return ret;
437 }
438 
439 /**
440  * p9_write_work - called when a transport can send some data
441  * @work: container for work to be done
442  *
443  */
444 
445 static void p9_write_work(struct work_struct *work)
446 {
447 	int n, err;
448 	struct p9_conn *m;
449 	struct p9_req_t *req;
450 
451 	m = container_of(work, struct p9_conn, wq);
452 
453 	if (m->err < 0) {
454 		clear_bit(Wworksched, &m->wsched);
455 		return;
456 	}
457 
458 	if (!m->wsize) {
459 		spin_lock(&m->client->lock);
460 		if (list_empty(&m->unsent_req_list)) {
461 			clear_bit(Wworksched, &m->wsched);
462 			spin_unlock(&m->client->lock);
463 			return;
464 		}
465 
466 		req = list_entry(m->unsent_req_list.next, struct p9_req_t,
467 			       req_list);
468 		req->status = REQ_STATUS_SENT;
469 		p9_debug(P9_DEBUG_TRANS, "move req %p\n", req);
470 		list_move_tail(&req->req_list, &m->req_list);
471 
472 		m->wbuf = req->tc->sdata;
473 		m->wsize = req->tc->size;
474 		m->wpos = 0;
475 		spin_unlock(&m->client->lock);
476 	}
477 
478 	p9_debug(P9_DEBUG_TRANS, "mux %p pos %d size %d\n",
479 		 m, m->wpos, m->wsize);
480 	clear_bit(Wpending, &m->wsched);
481 	err = p9_fd_write(m->client, m->wbuf + m->wpos, m->wsize - m->wpos);
482 	p9_debug(P9_DEBUG_TRANS, "mux %p sent %d bytes\n", m, err);
483 	if (err == -EAGAIN)
484 		goto end_clear;
485 
486 
487 	if (err < 0)
488 		goto error;
489 	else if (err == 0) {
490 		err = -EREMOTEIO;
491 		goto error;
492 	}
493 
494 	m->wpos += err;
495 	if (m->wpos == m->wsize)
496 		m->wpos = m->wsize = 0;
497 
498 end_clear:
499 	clear_bit(Wworksched, &m->wsched);
500 
501 	if (m->wsize || !list_empty(&m->unsent_req_list)) {
502 		if (test_and_clear_bit(Wpending, &m->wsched))
503 			n = POLLOUT;
504 		else
505 			n = p9_fd_poll(m->client, NULL);
506 
507 		if ((n & POLLOUT) &&
508 		   !test_and_set_bit(Wworksched, &m->wsched)) {
509 			p9_debug(P9_DEBUG_TRANS, "sched write work %p\n", m);
510 			schedule_work(&m->wq);
511 		}
512 	}
513 
514 	return;
515 
516 error:
517 	p9_conn_cancel(m, err);
518 	clear_bit(Wworksched, &m->wsched);
519 }
520 
521 static int p9_pollwake(wait_queue_t *wait, unsigned int mode, int sync, void *key)
522 {
523 	struct p9_poll_wait *pwait =
524 		container_of(wait, struct p9_poll_wait, wait);
525 	struct p9_conn *m = pwait->conn;
526 	unsigned long flags;
527 
528 	spin_lock_irqsave(&p9_poll_lock, flags);
529 	if (list_empty(&m->poll_pending_link))
530 		list_add_tail(&m->poll_pending_link, &p9_poll_pending_list);
531 	spin_unlock_irqrestore(&p9_poll_lock, flags);
532 
533 	schedule_work(&p9_poll_work);
534 	return 1;
535 }
536 
537 /**
538  * p9_pollwait - add poll task to the wait queue
539  * @filp: file pointer being polled
540  * @wait_address: wait_q to block on
541  * @p: poll state
542  *
543  * called by files poll operation to add v9fs-poll task to files wait queue
544  */
545 
546 static void
547 p9_pollwait(struct file *filp, wait_queue_head_t *wait_address, poll_table *p)
548 {
549 	struct p9_conn *m = container_of(p, struct p9_conn, pt);
550 	struct p9_poll_wait *pwait = NULL;
551 	int i;
552 
553 	for (i = 0; i < ARRAY_SIZE(m->poll_wait); i++) {
554 		if (m->poll_wait[i].wait_addr == NULL) {
555 			pwait = &m->poll_wait[i];
556 			break;
557 		}
558 	}
559 
560 	if (!pwait) {
561 		p9_debug(P9_DEBUG_ERROR, "not enough wait_address slots\n");
562 		return;
563 	}
564 
565 	pwait->conn = m;
566 	pwait->wait_addr = wait_address;
567 	init_waitqueue_func_entry(&pwait->wait, p9_pollwake);
568 	add_wait_queue(wait_address, &pwait->wait);
569 }
570 
571 /**
572  * p9_conn_create - initialize the per-session mux data
573  * @client: client instance
574  *
575  * Note: Creates the polling task if this is the first session.
576  */
577 
578 static void p9_conn_create(struct p9_client *client)
579 {
580 	int n;
581 	struct p9_trans_fd *ts = client->trans;
582 	struct p9_conn *m = &ts->conn;
583 
584 	p9_debug(P9_DEBUG_TRANS, "client %p msize %d\n", client, client->msize);
585 
586 	INIT_LIST_HEAD(&m->mux_list);
587 	m->client = client;
588 
589 	INIT_LIST_HEAD(&m->req_list);
590 	INIT_LIST_HEAD(&m->unsent_req_list);
591 	INIT_WORK(&m->rq, p9_read_work);
592 	INIT_WORK(&m->wq, p9_write_work);
593 	INIT_LIST_HEAD(&m->poll_pending_link);
594 	init_poll_funcptr(&m->pt, p9_pollwait);
595 
596 	n = p9_fd_poll(client, &m->pt);
597 	if (n & POLLIN) {
598 		p9_debug(P9_DEBUG_TRANS, "mux %p can read\n", m);
599 		set_bit(Rpending, &m->wsched);
600 	}
601 
602 	if (n & POLLOUT) {
603 		p9_debug(P9_DEBUG_TRANS, "mux %p can write\n", m);
604 		set_bit(Wpending, &m->wsched);
605 	}
606 }
607 
608 /**
609  * p9_poll_mux - polls a mux and schedules read or write works if necessary
610  * @m: connection to poll
611  *
612  */
613 
614 static void p9_poll_mux(struct p9_conn *m)
615 {
616 	int n;
617 
618 	if (m->err < 0)
619 		return;
620 
621 	n = p9_fd_poll(m->client, NULL);
622 	if (n < 0 || n & (POLLERR | POLLHUP | POLLNVAL)) {
623 		p9_debug(P9_DEBUG_TRANS, "error mux %p err %d\n", m, n);
624 		if (n >= 0)
625 			n = -ECONNRESET;
626 		p9_conn_cancel(m, n);
627 	}
628 
629 	if (n & POLLIN) {
630 		set_bit(Rpending, &m->wsched);
631 		p9_debug(P9_DEBUG_TRANS, "mux %p can read\n", m);
632 		if (!test_and_set_bit(Rworksched, &m->wsched)) {
633 			p9_debug(P9_DEBUG_TRANS, "sched read work %p\n", m);
634 			schedule_work(&m->rq);
635 		}
636 	}
637 
638 	if (n & POLLOUT) {
639 		set_bit(Wpending, &m->wsched);
640 		p9_debug(P9_DEBUG_TRANS, "mux %p can write\n", m);
641 		if ((m->wsize || !list_empty(&m->unsent_req_list)) &&
642 		    !test_and_set_bit(Wworksched, &m->wsched)) {
643 			p9_debug(P9_DEBUG_TRANS, "sched write work %p\n", m);
644 			schedule_work(&m->wq);
645 		}
646 	}
647 }
648 
649 /**
650  * p9_fd_request - send 9P request
651  * The function can sleep until the request is scheduled for sending.
652  * The function can be interrupted. Return from the function is not
653  * a guarantee that the request is sent successfully.
654  *
655  * @client: client instance
656  * @req: request to be sent
657  *
658  */
659 
660 static int p9_fd_request(struct p9_client *client, struct p9_req_t *req)
661 {
662 	int n;
663 	struct p9_trans_fd *ts = client->trans;
664 	struct p9_conn *m = &ts->conn;
665 
666 	p9_debug(P9_DEBUG_TRANS, "mux %p task %p tcall %p id %d\n",
667 		 m, current, req->tc, req->tc->id);
668 	if (m->err < 0)
669 		return m->err;
670 
671 	spin_lock(&client->lock);
672 	req->status = REQ_STATUS_UNSENT;
673 	list_add_tail(&req->req_list, &m->unsent_req_list);
674 	spin_unlock(&client->lock);
675 
676 	if (test_and_clear_bit(Wpending, &m->wsched))
677 		n = POLLOUT;
678 	else
679 		n = p9_fd_poll(m->client, NULL);
680 
681 	if (n & POLLOUT && !test_and_set_bit(Wworksched, &m->wsched))
682 		schedule_work(&m->wq);
683 
684 	return 0;
685 }
686 
687 static int p9_fd_cancel(struct p9_client *client, struct p9_req_t *req)
688 {
689 	int ret = 1;
690 
691 	p9_debug(P9_DEBUG_TRANS, "client %p req %p\n", client, req);
692 
693 	spin_lock(&client->lock);
694 
695 	if (req->status == REQ_STATUS_UNSENT) {
696 		list_del(&req->req_list);
697 		req->status = REQ_STATUS_FLSHD;
698 		ret = 0;
699 	}
700 	spin_unlock(&client->lock);
701 
702 	return ret;
703 }
704 
705 static int p9_fd_cancelled(struct p9_client *client, struct p9_req_t *req)
706 {
707 	p9_debug(P9_DEBUG_TRANS, "client %p req %p\n", client, req);
708 
709 	/* we haven't received a response for oldreq,
710 	 * remove it from the list.
711 	 */
712 	spin_lock(&client->lock);
713 	list_del(&req->req_list);
714 	spin_unlock(&client->lock);
715 
716 	return 0;
717 }
718 
719 /**
720  * parse_opts - parse mount options into p9_fd_opts structure
721  * @params: options string passed from mount
722  * @opts: fd transport-specific structure to parse options into
723  *
724  * Returns 0 upon success, -ERRNO upon failure
725  */
726 
727 static int parse_opts(char *params, struct p9_fd_opts *opts)
728 {
729 	char *p;
730 	substring_t args[MAX_OPT_ARGS];
731 	int option;
732 	char *options, *tmp_options;
733 
734 	opts->port = P9_PORT;
735 	opts->rfd = ~0;
736 	opts->wfd = ~0;
737 	opts->privport = 0;
738 
739 	if (!params)
740 		return 0;
741 
742 	tmp_options = kstrdup(params, GFP_KERNEL);
743 	if (!tmp_options) {
744 		p9_debug(P9_DEBUG_ERROR,
745 			 "failed to allocate copy of option string\n");
746 		return -ENOMEM;
747 	}
748 	options = tmp_options;
749 
750 	while ((p = strsep(&options, ",")) != NULL) {
751 		int token;
752 		int r;
753 		if (!*p)
754 			continue;
755 		token = match_token(p, tokens, args);
756 		if ((token != Opt_err) && (token != Opt_privport)) {
757 			r = match_int(&args[0], &option);
758 			if (r < 0) {
759 				p9_debug(P9_DEBUG_ERROR,
760 					 "integer field, but no integer?\n");
761 				continue;
762 			}
763 		}
764 		switch (token) {
765 		case Opt_port:
766 			opts->port = option;
767 			break;
768 		case Opt_rfdno:
769 			opts->rfd = option;
770 			break;
771 		case Opt_wfdno:
772 			opts->wfd = option;
773 			break;
774 		case Opt_privport:
775 			opts->privport = 1;
776 			break;
777 		default:
778 			continue;
779 		}
780 	}
781 
782 	kfree(tmp_options);
783 	return 0;
784 }
785 
786 static int p9_fd_open(struct p9_client *client, int rfd, int wfd)
787 {
788 	struct p9_trans_fd *ts = kzalloc(sizeof(struct p9_trans_fd),
789 					   GFP_KERNEL);
790 	if (!ts)
791 		return -ENOMEM;
792 
793 	ts->rd = fget(rfd);
794 	ts->wr = fget(wfd);
795 	if (!ts->rd || !ts->wr) {
796 		if (ts->rd)
797 			fput(ts->rd);
798 		if (ts->wr)
799 			fput(ts->wr);
800 		kfree(ts);
801 		return -EIO;
802 	}
803 
804 	client->trans = ts;
805 	client->status = Connected;
806 
807 	return 0;
808 }
809 
810 static int p9_socket_open(struct p9_client *client, struct socket *csocket)
811 {
812 	struct p9_trans_fd *p;
813 	struct file *file;
814 
815 	p = kzalloc(sizeof(struct p9_trans_fd), GFP_KERNEL);
816 	if (!p)
817 		return -ENOMEM;
818 
819 	csocket->sk->sk_allocation = GFP_NOIO;
820 	file = sock_alloc_file(csocket, 0, NULL);
821 	if (IS_ERR(file)) {
822 		pr_err("%s (%d): failed to map fd\n",
823 		       __func__, task_pid_nr(current));
824 		sock_release(csocket);
825 		kfree(p);
826 		return PTR_ERR(file);
827 	}
828 
829 	get_file(file);
830 	p->wr = p->rd = file;
831 	client->trans = p;
832 	client->status = Connected;
833 
834 	p->rd->f_flags |= O_NONBLOCK;
835 
836 	p9_conn_create(client);
837 	return 0;
838 }
839 
840 /**
841  * p9_mux_destroy - cancels all pending requests of mux
842  * @m: mux to destroy
843  *
844  */
845 
846 static void p9_conn_destroy(struct p9_conn *m)
847 {
848 	p9_debug(P9_DEBUG_TRANS, "mux %p prev %p next %p\n",
849 		 m, m->mux_list.prev, m->mux_list.next);
850 
851 	p9_mux_poll_stop(m);
852 	cancel_work_sync(&m->rq);
853 	cancel_work_sync(&m->wq);
854 
855 	p9_conn_cancel(m, -ECONNRESET);
856 
857 	m->client = NULL;
858 }
859 
860 /**
861  * p9_fd_close - shutdown file descriptor transport
862  * @client: client instance
863  *
864  */
865 
866 static void p9_fd_close(struct p9_client *client)
867 {
868 	struct p9_trans_fd *ts;
869 
870 	if (!client)
871 		return;
872 
873 	ts = client->trans;
874 	if (!ts)
875 		return;
876 
877 	client->status = Disconnected;
878 
879 	p9_conn_destroy(&ts->conn);
880 
881 	if (ts->rd)
882 		fput(ts->rd);
883 	if (ts->wr)
884 		fput(ts->wr);
885 
886 	kfree(ts);
887 }
888 
889 /*
890  * stolen from NFS - maybe should be made a generic function?
891  */
892 static inline int valid_ipaddr4(const char *buf)
893 {
894 	int rc, count, in[4];
895 
896 	rc = sscanf(buf, "%d.%d.%d.%d", &in[0], &in[1], &in[2], &in[3]);
897 	if (rc != 4)
898 		return -EINVAL;
899 	for (count = 0; count < 4; count++) {
900 		if (in[count] > 255)
901 			return -EINVAL;
902 	}
903 	return 0;
904 }
905 
906 static int p9_bind_privport(struct socket *sock)
907 {
908 	struct sockaddr_in cl;
909 	int port, err = -EINVAL;
910 
911 	memset(&cl, 0, sizeof(cl));
912 	cl.sin_family = AF_INET;
913 	cl.sin_addr.s_addr = INADDR_ANY;
914 	for (port = p9_ipport_resv_max; port >= p9_ipport_resv_min; port--) {
915 		cl.sin_port = htons((ushort)port);
916 		err = kernel_bind(sock, (struct sockaddr *)&cl, sizeof(cl));
917 		if (err != -EADDRINUSE)
918 			break;
919 	}
920 	return err;
921 }
922 
923 
924 static int
925 p9_fd_create_tcp(struct p9_client *client, const char *addr, char *args)
926 {
927 	int err;
928 	struct socket *csocket;
929 	struct sockaddr_in sin_server;
930 	struct p9_fd_opts opts;
931 
932 	err = parse_opts(args, &opts);
933 	if (err < 0)
934 		return err;
935 
936 	if (valid_ipaddr4(addr) < 0)
937 		return -EINVAL;
938 
939 	csocket = NULL;
940 
941 	sin_server.sin_family = AF_INET;
942 	sin_server.sin_addr.s_addr = in_aton(addr);
943 	sin_server.sin_port = htons(opts.port);
944 	err = __sock_create(current->nsproxy->net_ns, PF_INET,
945 			    SOCK_STREAM, IPPROTO_TCP, &csocket, 1);
946 	if (err) {
947 		pr_err("%s (%d): problem creating socket\n",
948 		       __func__, task_pid_nr(current));
949 		return err;
950 	}
951 
952 	if (opts.privport) {
953 		err = p9_bind_privport(csocket);
954 		if (err < 0) {
955 			pr_err("%s (%d): problem binding to privport\n",
956 			       __func__, task_pid_nr(current));
957 			sock_release(csocket);
958 			return err;
959 		}
960 	}
961 
962 	err = csocket->ops->connect(csocket,
963 				    (struct sockaddr *)&sin_server,
964 				    sizeof(struct sockaddr_in), 0);
965 	if (err < 0) {
966 		pr_err("%s (%d): problem connecting socket to %s\n",
967 		       __func__, task_pid_nr(current), addr);
968 		sock_release(csocket);
969 		return err;
970 	}
971 
972 	return p9_socket_open(client, csocket);
973 }
974 
975 static int
976 p9_fd_create_unix(struct p9_client *client, const char *addr, char *args)
977 {
978 	int err;
979 	struct socket *csocket;
980 	struct sockaddr_un sun_server;
981 
982 	csocket = NULL;
983 
984 	if (strlen(addr) >= UNIX_PATH_MAX) {
985 		pr_err("%s (%d): address too long: %s\n",
986 		       __func__, task_pid_nr(current), addr);
987 		return -ENAMETOOLONG;
988 	}
989 
990 	sun_server.sun_family = PF_UNIX;
991 	strcpy(sun_server.sun_path, addr);
992 	err = __sock_create(current->nsproxy->net_ns, PF_UNIX,
993 			    SOCK_STREAM, 0, &csocket, 1);
994 	if (err < 0) {
995 		pr_err("%s (%d): problem creating socket\n",
996 		       __func__, task_pid_nr(current));
997 
998 		return err;
999 	}
1000 	err = csocket->ops->connect(csocket, (struct sockaddr *)&sun_server,
1001 			sizeof(struct sockaddr_un) - 1, 0);
1002 	if (err < 0) {
1003 		pr_err("%s (%d): problem connecting socket: %s: %d\n",
1004 		       __func__, task_pid_nr(current), addr, err);
1005 		sock_release(csocket);
1006 		return err;
1007 	}
1008 
1009 	return p9_socket_open(client, csocket);
1010 }
1011 
1012 static int
1013 p9_fd_create(struct p9_client *client, const char *addr, char *args)
1014 {
1015 	int err;
1016 	struct p9_fd_opts opts;
1017 
1018 	parse_opts(args, &opts);
1019 
1020 	if (opts.rfd == ~0 || opts.wfd == ~0) {
1021 		pr_err("Insufficient options for proto=fd\n");
1022 		return -ENOPROTOOPT;
1023 	}
1024 
1025 	err = p9_fd_open(client, opts.rfd, opts.wfd);
1026 	if (err < 0)
1027 		return err;
1028 
1029 	p9_conn_create(client);
1030 
1031 	return 0;
1032 }
1033 
1034 static struct p9_trans_module p9_tcp_trans = {
1035 	.name = "tcp",
1036 	.maxsize = MAX_SOCK_BUF,
1037 	.def = 0,
1038 	.create = p9_fd_create_tcp,
1039 	.close = p9_fd_close,
1040 	.request = p9_fd_request,
1041 	.cancel = p9_fd_cancel,
1042 	.cancelled = p9_fd_cancelled,
1043 	.owner = THIS_MODULE,
1044 };
1045 
1046 static struct p9_trans_module p9_unix_trans = {
1047 	.name = "unix",
1048 	.maxsize = MAX_SOCK_BUF,
1049 	.def = 0,
1050 	.create = p9_fd_create_unix,
1051 	.close = p9_fd_close,
1052 	.request = p9_fd_request,
1053 	.cancel = p9_fd_cancel,
1054 	.cancelled = p9_fd_cancelled,
1055 	.owner = THIS_MODULE,
1056 };
1057 
1058 static struct p9_trans_module p9_fd_trans = {
1059 	.name = "fd",
1060 	.maxsize = MAX_SOCK_BUF,
1061 	.def = 0,
1062 	.create = p9_fd_create,
1063 	.close = p9_fd_close,
1064 	.request = p9_fd_request,
1065 	.cancel = p9_fd_cancel,
1066 	.cancelled = p9_fd_cancelled,
1067 	.owner = THIS_MODULE,
1068 };
1069 
1070 /**
1071  * p9_poll_proc - poll worker thread
1072  * @a: thread state and arguments
1073  *
1074  * polls all v9fs transports for new events and queues the appropriate
1075  * work to the work queue
1076  *
1077  */
1078 
1079 static void p9_poll_workfn(struct work_struct *work)
1080 {
1081 	unsigned long flags;
1082 
1083 	p9_debug(P9_DEBUG_TRANS, "start %p\n", current);
1084 
1085 	spin_lock_irqsave(&p9_poll_lock, flags);
1086 	while (!list_empty(&p9_poll_pending_list)) {
1087 		struct p9_conn *conn = list_first_entry(&p9_poll_pending_list,
1088 							struct p9_conn,
1089 							poll_pending_link);
1090 		list_del_init(&conn->poll_pending_link);
1091 		spin_unlock_irqrestore(&p9_poll_lock, flags);
1092 
1093 		p9_poll_mux(conn);
1094 
1095 		spin_lock_irqsave(&p9_poll_lock, flags);
1096 	}
1097 	spin_unlock_irqrestore(&p9_poll_lock, flags);
1098 
1099 	p9_debug(P9_DEBUG_TRANS, "finish\n");
1100 }
1101 
1102 int p9_trans_fd_init(void)
1103 {
1104 	v9fs_register_trans(&p9_tcp_trans);
1105 	v9fs_register_trans(&p9_unix_trans);
1106 	v9fs_register_trans(&p9_fd_trans);
1107 
1108 	return 0;
1109 }
1110 
1111 void p9_trans_fd_exit(void)
1112 {
1113 	flush_work(&p9_poll_work);
1114 	v9fs_unregister_trans(&p9_tcp_trans);
1115 	v9fs_unregister_trans(&p9_unix_trans);
1116 	v9fs_unregister_trans(&p9_fd_trans);
1117 }
1118