xref: /openbmc/linux/net/rose/af_rose.c (revision 96de0e252cedffad61b3cb5e05662c591898e69a)
1 /*
2  * This program is free software; you can redistribute it and/or modify
3  * it under the terms of the GNU General Public License as published by
4  * the Free Software Foundation; either version 2 of the License, or
5  * (at your option) any later version.
6  *
7  * Copyright (C) Jonathan Naylor G4KLX (g4klx@g4klx.demon.co.uk)
8  * Copyright (C) Alan Cox GW4PTS (alan@lxorguk.ukuu.org.uk)
9  * Copyright (C) Terry Dawson VK2KTJ (terry@animats.net)
10  * Copyright (C) Tomi Manninen OH2BNS (oh2bns@sral.fi)
11  */
12 
13 #include <linux/capability.h>
14 #include <linux/module.h>
15 #include <linux/moduleparam.h>
16 #include <linux/init.h>
17 #include <linux/errno.h>
18 #include <linux/types.h>
19 #include <linux/socket.h>
20 #include <linux/in.h>
21 #include <linux/kernel.h>
22 #include <linux/sched.h>
23 #include <linux/spinlock.h>
24 #include <linux/timer.h>
25 #include <linux/string.h>
26 #include <linux/sockios.h>
27 #include <linux/net.h>
28 #include <linux/stat.h>
29 #include <net/net_namespace.h>
30 #include <net/ax25.h>
31 #include <linux/inet.h>
32 #include <linux/netdevice.h>
33 #include <linux/if_arp.h>
34 #include <linux/skbuff.h>
35 #include <net/sock.h>
36 #include <asm/system.h>
37 #include <asm/uaccess.h>
38 #include <linux/fcntl.h>
39 #include <linux/termios.h>
40 #include <linux/mm.h>
41 #include <linux/interrupt.h>
42 #include <linux/notifier.h>
43 #include <net/rose.h>
44 #include <linux/proc_fs.h>
45 #include <linux/seq_file.h>
46 #include <net/tcp_states.h>
47 #include <net/ip.h>
48 #include <net/arp.h>
49 
50 static int rose_ndevs = 10;
51 
52 int sysctl_rose_restart_request_timeout = ROSE_DEFAULT_T0;
53 int sysctl_rose_call_request_timeout    = ROSE_DEFAULT_T1;
54 int sysctl_rose_reset_request_timeout   = ROSE_DEFAULT_T2;
55 int sysctl_rose_clear_request_timeout   = ROSE_DEFAULT_T3;
56 int sysctl_rose_no_activity_timeout     = ROSE_DEFAULT_IDLE;
57 int sysctl_rose_ack_hold_back_timeout   = ROSE_DEFAULT_HB;
58 int sysctl_rose_routing_control         = ROSE_DEFAULT_ROUTING;
59 int sysctl_rose_link_fail_timeout       = ROSE_DEFAULT_FAIL_TIMEOUT;
60 int sysctl_rose_maximum_vcs             = ROSE_DEFAULT_MAXVC;
61 int sysctl_rose_window_size             = ROSE_DEFAULT_WINDOW_SIZE;
62 
63 static HLIST_HEAD(rose_list);
64 static DEFINE_SPINLOCK(rose_list_lock);
65 
66 static struct proto_ops rose_proto_ops;
67 
68 ax25_address rose_callsign;
69 
70 /*
71  * ROSE network devices are virtual network devices encapsulating ROSE
72  * frames into AX.25 which will be sent through an AX.25 device, so form a
73  * special "super class" of normal net devices; split their locks off into a
74  * separate class since they always nest.
75  */
76 static struct lock_class_key rose_netdev_xmit_lock_key;
77 
78 /*
79  *	Convert a ROSE address into text.
80  */
81 const char *rose2asc(const rose_address *addr)
82 {
83 	static char buffer[11];
84 
85 	if (addr->rose_addr[0] == 0x00 && addr->rose_addr[1] == 0x00 &&
86 	    addr->rose_addr[2] == 0x00 && addr->rose_addr[3] == 0x00 &&
87 	    addr->rose_addr[4] == 0x00) {
88 		strcpy(buffer, "*");
89 	} else {
90 		sprintf(buffer, "%02X%02X%02X%02X%02X", addr->rose_addr[0] & 0xFF,
91 						addr->rose_addr[1] & 0xFF,
92 						addr->rose_addr[2] & 0xFF,
93 						addr->rose_addr[3] & 0xFF,
94 						addr->rose_addr[4] & 0xFF);
95 	}
96 
97 	return buffer;
98 }
99 
100 /*
101  *	Compare two ROSE addresses, 0 == equal.
102  */
103 int rosecmp(rose_address *addr1, rose_address *addr2)
104 {
105 	int i;
106 
107 	for (i = 0; i < 5; i++)
108 		if (addr1->rose_addr[i] != addr2->rose_addr[i])
109 			return 1;
110 
111 	return 0;
112 }
113 
114 /*
115  *	Compare two ROSE addresses for only mask digits, 0 == equal.
116  */
117 int rosecmpm(rose_address *addr1, rose_address *addr2, unsigned short mask)
118 {
119 	int i, j;
120 
121 	if (mask > 10)
122 		return 1;
123 
124 	for (i = 0; i < mask; i++) {
125 		j = i / 2;
126 
127 		if ((i % 2) != 0) {
128 			if ((addr1->rose_addr[j] & 0x0F) != (addr2->rose_addr[j] & 0x0F))
129 				return 1;
130 		} else {
131 			if ((addr1->rose_addr[j] & 0xF0) != (addr2->rose_addr[j] & 0xF0))
132 				return 1;
133 		}
134 	}
135 
136 	return 0;
137 }
138 
139 /*
140  *	Socket removal during an interrupt is now safe.
141  */
142 static void rose_remove_socket(struct sock *sk)
143 {
144 	spin_lock_bh(&rose_list_lock);
145 	sk_del_node_init(sk);
146 	spin_unlock_bh(&rose_list_lock);
147 }
148 
149 /*
150  *	Kill all bound sockets on a broken link layer connection to a
151  *	particular neighbour.
152  */
153 void rose_kill_by_neigh(struct rose_neigh *neigh)
154 {
155 	struct sock *s;
156 	struct hlist_node *node;
157 
158 	spin_lock_bh(&rose_list_lock);
159 	sk_for_each(s, node, &rose_list) {
160 		struct rose_sock *rose = rose_sk(s);
161 
162 		if (rose->neighbour == neigh) {
163 			rose_disconnect(s, ENETUNREACH, ROSE_OUT_OF_ORDER, 0);
164 			rose->neighbour->use--;
165 			rose->neighbour = NULL;
166 		}
167 	}
168 	spin_unlock_bh(&rose_list_lock);
169 }
170 
171 /*
172  *	Kill all bound sockets on a dropped device.
173  */
174 static void rose_kill_by_device(struct net_device *dev)
175 {
176 	struct sock *s;
177 	struct hlist_node *node;
178 
179 	spin_lock_bh(&rose_list_lock);
180 	sk_for_each(s, node, &rose_list) {
181 		struct rose_sock *rose = rose_sk(s);
182 
183 		if (rose->device == dev) {
184 			rose_disconnect(s, ENETUNREACH, ROSE_OUT_OF_ORDER, 0);
185 			rose->neighbour->use--;
186 			rose->device = NULL;
187 		}
188 	}
189 	spin_unlock_bh(&rose_list_lock);
190 }
191 
192 /*
193  *	Handle device status changes.
194  */
195 static int rose_device_event(struct notifier_block *this, unsigned long event,
196 	void *ptr)
197 {
198 	struct net_device *dev = (struct net_device *)ptr;
199 
200 	if (dev->nd_net != &init_net)
201 		return NOTIFY_DONE;
202 
203 	if (event != NETDEV_DOWN)
204 		return NOTIFY_DONE;
205 
206 	switch (dev->type) {
207 	case ARPHRD_ROSE:
208 		rose_kill_by_device(dev);
209 		break;
210 	case ARPHRD_AX25:
211 		rose_link_device_down(dev);
212 		rose_rt_device_down(dev);
213 		break;
214 	}
215 
216 	return NOTIFY_DONE;
217 }
218 
219 /*
220  *	Add a socket to the bound sockets list.
221  */
222 static void rose_insert_socket(struct sock *sk)
223 {
224 
225 	spin_lock_bh(&rose_list_lock);
226 	sk_add_node(sk, &rose_list);
227 	spin_unlock_bh(&rose_list_lock);
228 }
229 
230 /*
231  *	Find a socket that wants to accept the Call Request we just
232  *	received.
233  */
234 static struct sock *rose_find_listener(rose_address *addr, ax25_address *call)
235 {
236 	struct sock *s;
237 	struct hlist_node *node;
238 
239 	spin_lock_bh(&rose_list_lock);
240 	sk_for_each(s, node, &rose_list) {
241 		struct rose_sock *rose = rose_sk(s);
242 
243 		if (!rosecmp(&rose->source_addr, addr) &&
244 		    !ax25cmp(&rose->source_call, call) &&
245 		    !rose->source_ndigis && s->sk_state == TCP_LISTEN)
246 			goto found;
247 	}
248 
249 	sk_for_each(s, node, &rose_list) {
250 		struct rose_sock *rose = rose_sk(s);
251 
252 		if (!rosecmp(&rose->source_addr, addr) &&
253 		    !ax25cmp(&rose->source_call, &null_ax25_address) &&
254 		    s->sk_state == TCP_LISTEN)
255 			goto found;
256 	}
257 	s = NULL;
258 found:
259 	spin_unlock_bh(&rose_list_lock);
260 	return s;
261 }
262 
263 /*
264  *	Find a connected ROSE socket given my LCI and device.
265  */
266 struct sock *rose_find_socket(unsigned int lci, struct rose_neigh *neigh)
267 {
268 	struct sock *s;
269 	struct hlist_node *node;
270 
271 	spin_lock_bh(&rose_list_lock);
272 	sk_for_each(s, node, &rose_list) {
273 		struct rose_sock *rose = rose_sk(s);
274 
275 		if (rose->lci == lci && rose->neighbour == neigh)
276 			goto found;
277 	}
278 	s = NULL;
279 found:
280 	spin_unlock_bh(&rose_list_lock);
281 	return s;
282 }
283 
284 /*
285  *	Find a unique LCI for a given device.
286  */
287 unsigned int rose_new_lci(struct rose_neigh *neigh)
288 {
289 	int lci;
290 
291 	if (neigh->dce_mode) {
292 		for (lci = 1; lci <= sysctl_rose_maximum_vcs; lci++)
293 			if (rose_find_socket(lci, neigh) == NULL && rose_route_free_lci(lci, neigh) == NULL)
294 				return lci;
295 	} else {
296 		for (lci = sysctl_rose_maximum_vcs; lci > 0; lci--)
297 			if (rose_find_socket(lci, neigh) == NULL && rose_route_free_lci(lci, neigh) == NULL)
298 				return lci;
299 	}
300 
301 	return 0;
302 }
303 
304 /*
305  *	Deferred destroy.
306  */
307 void rose_destroy_socket(struct sock *);
308 
309 /*
310  *	Handler for deferred kills.
311  */
312 static void rose_destroy_timer(unsigned long data)
313 {
314 	rose_destroy_socket((struct sock *)data);
315 }
316 
317 /*
318  *	This is called from user mode and the timers. Thus it protects itself
319  *	against interrupt users but doesn't worry about being called during
320  *	work.  Once it is removed from the queue no interrupt or bottom half
321  *	will touch it and we are (fairly 8-) ) safe.
322  */
323 void rose_destroy_socket(struct sock *sk)
324 {
325 	struct sk_buff *skb;
326 
327 	rose_remove_socket(sk);
328 	rose_stop_heartbeat(sk);
329 	rose_stop_idletimer(sk);
330 	rose_stop_timer(sk);
331 
332 	rose_clear_queues(sk);		/* Flush the queues */
333 
334 	while ((skb = skb_dequeue(&sk->sk_receive_queue)) != NULL) {
335 		if (skb->sk != sk) {	/* A pending connection */
336 			/* Queue the unaccepted socket for death */
337 			sock_set_flag(skb->sk, SOCK_DEAD);
338 			rose_start_heartbeat(skb->sk);
339 			rose_sk(skb->sk)->state = ROSE_STATE_0;
340 		}
341 
342 		kfree_skb(skb);
343 	}
344 
345 	if (atomic_read(&sk->sk_wmem_alloc) ||
346 	    atomic_read(&sk->sk_rmem_alloc)) {
347 		/* Defer: outstanding buffers */
348 		init_timer(&sk->sk_timer);
349 		sk->sk_timer.expires  = jiffies + 10 * HZ;
350 		sk->sk_timer.function = rose_destroy_timer;
351 		sk->sk_timer.data     = (unsigned long)sk;
352 		add_timer(&sk->sk_timer);
353 	} else
354 		sock_put(sk);
355 }
356 
357 /*
358  *	Handling for system calls applied via the various interfaces to a
359  *	ROSE socket object.
360  */
361 
362 static int rose_setsockopt(struct socket *sock, int level, int optname,
363 	char __user *optval, int optlen)
364 {
365 	struct sock *sk = sock->sk;
366 	struct rose_sock *rose = rose_sk(sk);
367 	int opt;
368 
369 	if (level != SOL_ROSE)
370 		return -ENOPROTOOPT;
371 
372 	if (optlen < sizeof(int))
373 		return -EINVAL;
374 
375 	if (get_user(opt, (int __user *)optval))
376 		return -EFAULT;
377 
378 	switch (optname) {
379 	case ROSE_DEFER:
380 		rose->defer = opt ? 1 : 0;
381 		return 0;
382 
383 	case ROSE_T1:
384 		if (opt < 1)
385 			return -EINVAL;
386 		rose->t1 = opt * HZ;
387 		return 0;
388 
389 	case ROSE_T2:
390 		if (opt < 1)
391 			return -EINVAL;
392 		rose->t2 = opt * HZ;
393 		return 0;
394 
395 	case ROSE_T3:
396 		if (opt < 1)
397 			return -EINVAL;
398 		rose->t3 = opt * HZ;
399 		return 0;
400 
401 	case ROSE_HOLDBACK:
402 		if (opt < 1)
403 			return -EINVAL;
404 		rose->hb = opt * HZ;
405 		return 0;
406 
407 	case ROSE_IDLE:
408 		if (opt < 0)
409 			return -EINVAL;
410 		rose->idle = opt * 60 * HZ;
411 		return 0;
412 
413 	case ROSE_QBITINCL:
414 		rose->qbitincl = opt ? 1 : 0;
415 		return 0;
416 
417 	default:
418 		return -ENOPROTOOPT;
419 	}
420 }
421 
422 static int rose_getsockopt(struct socket *sock, int level, int optname,
423 	char __user *optval, int __user *optlen)
424 {
425 	struct sock *sk = sock->sk;
426 	struct rose_sock *rose = rose_sk(sk);
427 	int val = 0;
428 	int len;
429 
430 	if (level != SOL_ROSE)
431 		return -ENOPROTOOPT;
432 
433 	if (get_user(len, optlen))
434 		return -EFAULT;
435 
436 	if (len < 0)
437 		return -EINVAL;
438 
439 	switch (optname) {
440 	case ROSE_DEFER:
441 		val = rose->defer;
442 		break;
443 
444 	case ROSE_T1:
445 		val = rose->t1 / HZ;
446 		break;
447 
448 	case ROSE_T2:
449 		val = rose->t2 / HZ;
450 		break;
451 
452 	case ROSE_T3:
453 		val = rose->t3 / HZ;
454 		break;
455 
456 	case ROSE_HOLDBACK:
457 		val = rose->hb / HZ;
458 		break;
459 
460 	case ROSE_IDLE:
461 		val = rose->idle / (60 * HZ);
462 		break;
463 
464 	case ROSE_QBITINCL:
465 		val = rose->qbitincl;
466 		break;
467 
468 	default:
469 		return -ENOPROTOOPT;
470 	}
471 
472 	len = min_t(unsigned int, len, sizeof(int));
473 
474 	if (put_user(len, optlen))
475 		return -EFAULT;
476 
477 	return copy_to_user(optval, &val, len) ? -EFAULT : 0;
478 }
479 
480 static int rose_listen(struct socket *sock, int backlog)
481 {
482 	struct sock *sk = sock->sk;
483 
484 	if (sk->sk_state != TCP_LISTEN) {
485 		struct rose_sock *rose = rose_sk(sk);
486 
487 		rose->dest_ndigis = 0;
488 		memset(&rose->dest_addr, 0, ROSE_ADDR_LEN);
489 		memset(&rose->dest_call, 0, AX25_ADDR_LEN);
490 		memset(rose->dest_digis, 0, AX25_ADDR_LEN * ROSE_MAX_DIGIS);
491 		sk->sk_max_ack_backlog = backlog;
492 		sk->sk_state           = TCP_LISTEN;
493 		return 0;
494 	}
495 
496 	return -EOPNOTSUPP;
497 }
498 
499 static struct proto rose_proto = {
500 	.name	  = "ROSE",
501 	.owner	  = THIS_MODULE,
502 	.obj_size = sizeof(struct rose_sock),
503 };
504 
505 static int rose_create(struct net *net, struct socket *sock, int protocol)
506 {
507 	struct sock *sk;
508 	struct rose_sock *rose;
509 
510 	if (net != &init_net)
511 		return -EAFNOSUPPORT;
512 
513 	if (sock->type != SOCK_SEQPACKET || protocol != 0)
514 		return -ESOCKTNOSUPPORT;
515 
516 	if ((sk = sk_alloc(net, PF_ROSE, GFP_ATOMIC, &rose_proto, 1)) == NULL)
517 		return -ENOMEM;
518 
519 	rose = rose_sk(sk);
520 
521 	sock_init_data(sock, sk);
522 
523 	skb_queue_head_init(&rose->ack_queue);
524 #ifdef M_BIT
525 	skb_queue_head_init(&rose->frag_queue);
526 	rose->fraglen    = 0;
527 #endif
528 
529 	sock->ops    = &rose_proto_ops;
530 	sk->sk_protocol = protocol;
531 
532 	init_timer(&rose->timer);
533 	init_timer(&rose->idletimer);
534 
535 	rose->t1   = msecs_to_jiffies(sysctl_rose_call_request_timeout);
536 	rose->t2   = msecs_to_jiffies(sysctl_rose_reset_request_timeout);
537 	rose->t3   = msecs_to_jiffies(sysctl_rose_clear_request_timeout);
538 	rose->hb   = msecs_to_jiffies(sysctl_rose_ack_hold_back_timeout);
539 	rose->idle = msecs_to_jiffies(sysctl_rose_no_activity_timeout);
540 
541 	rose->state = ROSE_STATE_0;
542 
543 	return 0;
544 }
545 
546 static struct sock *rose_make_new(struct sock *osk)
547 {
548 	struct sock *sk;
549 	struct rose_sock *rose, *orose;
550 
551 	if (osk->sk_type != SOCK_SEQPACKET)
552 		return NULL;
553 
554 	if ((sk = sk_alloc(osk->sk_net, PF_ROSE, GFP_ATOMIC, &rose_proto, 1)) == NULL)
555 		return NULL;
556 
557 	rose = rose_sk(sk);
558 
559 	sock_init_data(NULL, sk);
560 
561 	skb_queue_head_init(&rose->ack_queue);
562 #ifdef M_BIT
563 	skb_queue_head_init(&rose->frag_queue);
564 	rose->fraglen  = 0;
565 #endif
566 
567 	sk->sk_type     = osk->sk_type;
568 	sk->sk_socket   = osk->sk_socket;
569 	sk->sk_priority = osk->sk_priority;
570 	sk->sk_protocol = osk->sk_protocol;
571 	sk->sk_rcvbuf   = osk->sk_rcvbuf;
572 	sk->sk_sndbuf   = osk->sk_sndbuf;
573 	sk->sk_state    = TCP_ESTABLISHED;
574 	sk->sk_sleep    = osk->sk_sleep;
575 	sock_copy_flags(sk, osk);
576 
577 	init_timer(&rose->timer);
578 	init_timer(&rose->idletimer);
579 
580 	orose		= rose_sk(osk);
581 	rose->t1	= orose->t1;
582 	rose->t2	= orose->t2;
583 	rose->t3	= orose->t3;
584 	rose->hb	= orose->hb;
585 	rose->idle	= orose->idle;
586 	rose->defer	= orose->defer;
587 	rose->device	= orose->device;
588 	rose->qbitincl	= orose->qbitincl;
589 
590 	return sk;
591 }
592 
593 static int rose_release(struct socket *sock)
594 {
595 	struct sock *sk = sock->sk;
596 	struct rose_sock *rose;
597 
598 	if (sk == NULL) return 0;
599 
600 	rose = rose_sk(sk);
601 
602 	switch (rose->state) {
603 	case ROSE_STATE_0:
604 		rose_disconnect(sk, 0, -1, -1);
605 		rose_destroy_socket(sk);
606 		break;
607 
608 	case ROSE_STATE_2:
609 		rose->neighbour->use--;
610 		rose_disconnect(sk, 0, -1, -1);
611 		rose_destroy_socket(sk);
612 		break;
613 
614 	case ROSE_STATE_1:
615 	case ROSE_STATE_3:
616 	case ROSE_STATE_4:
617 	case ROSE_STATE_5:
618 		rose_clear_queues(sk);
619 		rose_stop_idletimer(sk);
620 		rose_write_internal(sk, ROSE_CLEAR_REQUEST);
621 		rose_start_t3timer(sk);
622 		rose->state  = ROSE_STATE_2;
623 		sk->sk_state    = TCP_CLOSE;
624 		sk->sk_shutdown |= SEND_SHUTDOWN;
625 		sk->sk_state_change(sk);
626 		sock_set_flag(sk, SOCK_DEAD);
627 		sock_set_flag(sk, SOCK_DESTROY);
628 		break;
629 
630 	default:
631 		break;
632 	}
633 
634 	sock->sk = NULL;
635 
636 	return 0;
637 }
638 
639 static int rose_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len)
640 {
641 	struct sock *sk = sock->sk;
642 	struct rose_sock *rose = rose_sk(sk);
643 	struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr;
644 	struct net_device *dev;
645 	ax25_address *source;
646 	ax25_uid_assoc *user;
647 	int n;
648 
649 	if (!sock_flag(sk, SOCK_ZAPPED))
650 		return -EINVAL;
651 
652 	if (addr_len != sizeof(struct sockaddr_rose) && addr_len != sizeof(struct full_sockaddr_rose))
653 		return -EINVAL;
654 
655 	if (addr->srose_family != AF_ROSE)
656 		return -EINVAL;
657 
658 	if (addr_len == sizeof(struct sockaddr_rose) && addr->srose_ndigis > 1)
659 		return -EINVAL;
660 
661 	if (addr->srose_ndigis > ROSE_MAX_DIGIS)
662 		return -EINVAL;
663 
664 	if ((dev = rose_dev_get(&addr->srose_addr)) == NULL) {
665 		SOCK_DEBUG(sk, "ROSE: bind failed: invalid address\n");
666 		return -EADDRNOTAVAIL;
667 	}
668 
669 	source = &addr->srose_call;
670 
671 	user = ax25_findbyuid(current->euid);
672 	if (user) {
673 		rose->source_call = user->call;
674 		ax25_uid_put(user);
675 	} else {
676 		if (ax25_uid_policy && !capable(CAP_NET_BIND_SERVICE))
677 			return -EACCES;
678 		rose->source_call   = *source;
679 	}
680 
681 	rose->source_addr   = addr->srose_addr;
682 	rose->device        = dev;
683 	rose->source_ndigis = addr->srose_ndigis;
684 
685 	if (addr_len == sizeof(struct full_sockaddr_rose)) {
686 		struct full_sockaddr_rose *full_addr = (struct full_sockaddr_rose *)uaddr;
687 		for (n = 0 ; n < addr->srose_ndigis ; n++)
688 			rose->source_digis[n] = full_addr->srose_digis[n];
689 	} else {
690 		if (rose->source_ndigis == 1) {
691 			rose->source_digis[0] = addr->srose_digi;
692 		}
693 	}
694 
695 	rose_insert_socket(sk);
696 
697 	sock_reset_flag(sk, SOCK_ZAPPED);
698 	SOCK_DEBUG(sk, "ROSE: socket is bound\n");
699 	return 0;
700 }
701 
702 static int rose_connect(struct socket *sock, struct sockaddr *uaddr, int addr_len, int flags)
703 {
704 	struct sock *sk = sock->sk;
705 	struct rose_sock *rose = rose_sk(sk);
706 	struct sockaddr_rose *addr = (struct sockaddr_rose *)uaddr;
707 	unsigned char cause, diagnostic;
708 	struct net_device *dev;
709 	ax25_uid_assoc *user;
710 	int n, err = 0;
711 
712 	if (addr_len != sizeof(struct sockaddr_rose) && addr_len != sizeof(struct full_sockaddr_rose))
713 		return -EINVAL;
714 
715 	if (addr->srose_family != AF_ROSE)
716 		return -EINVAL;
717 
718 	if (addr_len == sizeof(struct sockaddr_rose) && addr->srose_ndigis > 1)
719 		return -EINVAL;
720 
721 	if (addr->srose_ndigis > ROSE_MAX_DIGIS)
722 		return -EINVAL;
723 
724 	/* Source + Destination digis should not exceed ROSE_MAX_DIGIS */
725 	if ((rose->source_ndigis + addr->srose_ndigis) > ROSE_MAX_DIGIS)
726 		return -EINVAL;
727 
728 	lock_sock(sk);
729 
730 	if (sk->sk_state == TCP_ESTABLISHED && sock->state == SS_CONNECTING) {
731 		/* Connect completed during a ERESTARTSYS event */
732 		sock->state = SS_CONNECTED;
733 		goto out_release;
734 	}
735 
736 	if (sk->sk_state == TCP_CLOSE && sock->state == SS_CONNECTING) {
737 		sock->state = SS_UNCONNECTED;
738 		err = -ECONNREFUSED;
739 		goto out_release;
740 	}
741 
742 	if (sk->sk_state == TCP_ESTABLISHED) {
743 		/* No reconnect on a seqpacket socket */
744 		err = -EISCONN;
745 		goto out_release;
746 	}
747 
748 	sk->sk_state   = TCP_CLOSE;
749 	sock->state = SS_UNCONNECTED;
750 
751 	rose->neighbour = rose_get_neigh(&addr->srose_addr, &cause,
752 					 &diagnostic);
753 	if (!rose->neighbour)
754 		return -ENETUNREACH;
755 
756 	rose->lci = rose_new_lci(rose->neighbour);
757 	if (!rose->lci) {
758 		err = -ENETUNREACH;
759 		goto out_release;
760 	}
761 
762 	if (sock_flag(sk, SOCK_ZAPPED)) {	/* Must bind first - autobinding in this may or may not work */
763 		sock_reset_flag(sk, SOCK_ZAPPED);
764 
765 		if ((dev = rose_dev_first()) == NULL) {
766 			err = -ENETUNREACH;
767 			goto out_release;
768 		}
769 
770 		user = ax25_findbyuid(current->euid);
771 		if (!user) {
772 			err = -EINVAL;
773 			goto out_release;
774 		}
775 
776 		memcpy(&rose->source_addr, dev->dev_addr, ROSE_ADDR_LEN);
777 		rose->source_call = user->call;
778 		rose->device      = dev;
779 		ax25_uid_put(user);
780 
781 		rose_insert_socket(sk);		/* Finish the bind */
782 	}
783 rose_try_next_neigh:
784 	rose->dest_addr   = addr->srose_addr;
785 	rose->dest_call   = addr->srose_call;
786 	rose->rand        = ((long)rose & 0xFFFF) + rose->lci;
787 	rose->dest_ndigis = addr->srose_ndigis;
788 
789 	if (addr_len == sizeof(struct full_sockaddr_rose)) {
790 		struct full_sockaddr_rose *full_addr = (struct full_sockaddr_rose *)uaddr;
791 		for (n = 0 ; n < addr->srose_ndigis ; n++)
792 			rose->dest_digis[n] = full_addr->srose_digis[n];
793 	} else {
794 		if (rose->dest_ndigis == 1) {
795 			rose->dest_digis[0] = addr->srose_digi;
796 		}
797 	}
798 
799 	/* Move to connecting socket, start sending Connect Requests */
800 	sock->state   = SS_CONNECTING;
801 	sk->sk_state     = TCP_SYN_SENT;
802 
803 	rose->state = ROSE_STATE_1;
804 
805 	rose->neighbour->use++;
806 
807 	rose_write_internal(sk, ROSE_CALL_REQUEST);
808 	rose_start_heartbeat(sk);
809 	rose_start_t1timer(sk);
810 
811 	/* Now the loop */
812 	if (sk->sk_state != TCP_ESTABLISHED && (flags & O_NONBLOCK)) {
813 		err = -EINPROGRESS;
814 		goto out_release;
815 	}
816 
817 	/*
818 	 * A Connect Ack with Choke or timeout or failed routing will go to
819 	 * closed.
820 	 */
821 	if (sk->sk_state == TCP_SYN_SENT) {
822 		DEFINE_WAIT(wait);
823 
824 		for (;;) {
825 			prepare_to_wait(sk->sk_sleep, &wait,
826 					TASK_INTERRUPTIBLE);
827 			if (sk->sk_state != TCP_SYN_SENT)
828 				break;
829 			if (!signal_pending(current)) {
830 				release_sock(sk);
831 				schedule();
832 				lock_sock(sk);
833 				continue;
834 			}
835 			err = -ERESTARTSYS;
836 			break;
837 		}
838 		finish_wait(sk->sk_sleep, &wait);
839 
840 		if (err)
841 			goto out_release;
842 	}
843 
844 	if (sk->sk_state != TCP_ESTABLISHED) {
845 	/* Try next neighbour */
846 		rose->neighbour = rose_get_neigh(&addr->srose_addr, &cause, &diagnostic);
847 		if (rose->neighbour)
848 			goto rose_try_next_neigh;
849 
850 		/* No more neighbours */
851 		sock->state = SS_UNCONNECTED;
852 		err = sock_error(sk);	/* Always set at this point */
853 		goto out_release;
854 	}
855 
856 	sock->state = SS_CONNECTED;
857 
858 out_release:
859 	release_sock(sk);
860 
861 	return err;
862 }
863 
864 static int rose_accept(struct socket *sock, struct socket *newsock, int flags)
865 {
866 	struct sk_buff *skb;
867 	struct sock *newsk;
868 	DEFINE_WAIT(wait);
869 	struct sock *sk;
870 	int err = 0;
871 
872 	if ((sk = sock->sk) == NULL)
873 		return -EINVAL;
874 
875 	lock_sock(sk);
876 	if (sk->sk_type != SOCK_SEQPACKET) {
877 		err = -EOPNOTSUPP;
878 		goto out_release;
879 	}
880 
881 	if (sk->sk_state != TCP_LISTEN) {
882 		err = -EINVAL;
883 		goto out_release;
884 	}
885 
886 	/*
887 	 *	The write queue this time is holding sockets ready to use
888 	 *	hooked into the SABM we saved
889 	 */
890 	for (;;) {
891 		prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
892 
893 		skb = skb_dequeue(&sk->sk_receive_queue);
894 		if (skb)
895 			break;
896 
897 		if (flags & O_NONBLOCK) {
898 			err = -EWOULDBLOCK;
899 			break;
900 		}
901 		if (!signal_pending(current)) {
902 			release_sock(sk);
903 			schedule();
904 			lock_sock(sk);
905 			continue;
906 		}
907 		err = -ERESTARTSYS;
908 		break;
909 	}
910 	finish_wait(sk->sk_sleep, &wait);
911 	if (err)
912 		goto out_release;
913 
914 	newsk = skb->sk;
915 	newsk->sk_socket = newsock;
916 	newsk->sk_sleep = &newsock->wait;
917 
918 	/* Now attach up the new socket */
919 	skb->sk = NULL;
920 	kfree_skb(skb);
921 	sk->sk_ack_backlog--;
922 	newsock->sk = newsk;
923 
924 out_release:
925 	release_sock(sk);
926 
927 	return err;
928 }
929 
930 static int rose_getname(struct socket *sock, struct sockaddr *uaddr,
931 	int *uaddr_len, int peer)
932 {
933 	struct full_sockaddr_rose *srose = (struct full_sockaddr_rose *)uaddr;
934 	struct sock *sk = sock->sk;
935 	struct rose_sock *rose = rose_sk(sk);
936 	int n;
937 
938 	if (peer != 0) {
939 		if (sk->sk_state != TCP_ESTABLISHED)
940 			return -ENOTCONN;
941 		srose->srose_family = AF_ROSE;
942 		srose->srose_addr   = rose->dest_addr;
943 		srose->srose_call   = rose->dest_call;
944 		srose->srose_ndigis = rose->dest_ndigis;
945 		for (n = 0; n < rose->dest_ndigis; n++)
946 			srose->srose_digis[n] = rose->dest_digis[n];
947 	} else {
948 		srose->srose_family = AF_ROSE;
949 		srose->srose_addr   = rose->source_addr;
950 		srose->srose_call   = rose->source_call;
951 		srose->srose_ndigis = rose->source_ndigis;
952 		for (n = 0; n < rose->source_ndigis; n++)
953 			srose->srose_digis[n] = rose->source_digis[n];
954 	}
955 
956 	*uaddr_len = sizeof(struct full_sockaddr_rose);
957 	return 0;
958 }
959 
960 int rose_rx_call_request(struct sk_buff *skb, struct net_device *dev, struct rose_neigh *neigh, unsigned int lci)
961 {
962 	struct sock *sk;
963 	struct sock *make;
964 	struct rose_sock *make_rose;
965 	struct rose_facilities_struct facilities;
966 	int n, len;
967 
968 	skb->sk = NULL;		/* Initially we don't know who it's for */
969 
970 	/*
971 	 *	skb->data points to the rose frame start
972 	 */
973 	memset(&facilities, 0x00, sizeof(struct rose_facilities_struct));
974 
975 	len  = (((skb->data[3] >> 4) & 0x0F) + 1) / 2;
976 	len += (((skb->data[3] >> 0) & 0x0F) + 1) / 2;
977 	if (!rose_parse_facilities(skb->data + len + 4, &facilities)) {
978 		rose_transmit_clear_request(neigh, lci, ROSE_INVALID_FACILITY, 76);
979 		return 0;
980 	}
981 
982 	sk = rose_find_listener(&facilities.source_addr, &facilities.source_call);
983 
984 	/*
985 	 * We can't accept the Call Request.
986 	 */
987 	if (sk == NULL || sk_acceptq_is_full(sk) ||
988 	    (make = rose_make_new(sk)) == NULL) {
989 		rose_transmit_clear_request(neigh, lci, ROSE_NETWORK_CONGESTION, 120);
990 		return 0;
991 	}
992 
993 	skb->sk     = make;
994 	make->sk_state = TCP_ESTABLISHED;
995 	make_rose = rose_sk(make);
996 
997 	make_rose->lci           = lci;
998 	make_rose->dest_addr     = facilities.dest_addr;
999 	make_rose->dest_call     = facilities.dest_call;
1000 	make_rose->dest_ndigis   = facilities.dest_ndigis;
1001 	for (n = 0 ; n < facilities.dest_ndigis ; n++)
1002 		make_rose->dest_digis[n] = facilities.dest_digis[n];
1003 	make_rose->source_addr   = facilities.source_addr;
1004 	make_rose->source_call   = facilities.source_call;
1005 	make_rose->source_ndigis = facilities.source_ndigis;
1006 	for (n = 0 ; n < facilities.source_ndigis ; n++)
1007 		make_rose->source_digis[n]= facilities.source_digis[n];
1008 	make_rose->neighbour     = neigh;
1009 	make_rose->device        = dev;
1010 	make_rose->facilities    = facilities;
1011 
1012 	make_rose->neighbour->use++;
1013 
1014 	if (rose_sk(sk)->defer) {
1015 		make_rose->state = ROSE_STATE_5;
1016 	} else {
1017 		rose_write_internal(make, ROSE_CALL_ACCEPTED);
1018 		make_rose->state = ROSE_STATE_3;
1019 		rose_start_idletimer(make);
1020 	}
1021 
1022 	make_rose->condition = 0x00;
1023 	make_rose->vs        = 0;
1024 	make_rose->va        = 0;
1025 	make_rose->vr        = 0;
1026 	make_rose->vl        = 0;
1027 	sk->sk_ack_backlog++;
1028 
1029 	rose_insert_socket(make);
1030 
1031 	skb_queue_head(&sk->sk_receive_queue, skb);
1032 
1033 	rose_start_heartbeat(make);
1034 
1035 	if (!sock_flag(sk, SOCK_DEAD))
1036 		sk->sk_data_ready(sk, skb->len);
1037 
1038 	return 1;
1039 }
1040 
1041 static int rose_sendmsg(struct kiocb *iocb, struct socket *sock,
1042 			struct msghdr *msg, size_t len)
1043 {
1044 	struct sock *sk = sock->sk;
1045 	struct rose_sock *rose = rose_sk(sk);
1046 	struct sockaddr_rose *usrose = (struct sockaddr_rose *)msg->msg_name;
1047 	int err;
1048 	struct full_sockaddr_rose srose;
1049 	struct sk_buff *skb;
1050 	unsigned char *asmptr;
1051 	int n, size, qbit = 0;
1052 
1053 	if (msg->msg_flags & ~(MSG_DONTWAIT|MSG_EOR|MSG_CMSG_COMPAT))
1054 		return -EINVAL;
1055 
1056 	if (sock_flag(sk, SOCK_ZAPPED))
1057 		return -EADDRNOTAVAIL;
1058 
1059 	if (sk->sk_shutdown & SEND_SHUTDOWN) {
1060 		send_sig(SIGPIPE, current, 0);
1061 		return -EPIPE;
1062 	}
1063 
1064 	if (rose->neighbour == NULL || rose->device == NULL)
1065 		return -ENETUNREACH;
1066 
1067 	if (usrose != NULL) {
1068 		if (msg->msg_namelen != sizeof(struct sockaddr_rose) && msg->msg_namelen != sizeof(struct full_sockaddr_rose))
1069 			return -EINVAL;
1070 		memset(&srose, 0, sizeof(struct full_sockaddr_rose));
1071 		memcpy(&srose, usrose, msg->msg_namelen);
1072 		if (rosecmp(&rose->dest_addr, &srose.srose_addr) != 0 ||
1073 		    ax25cmp(&rose->dest_call, &srose.srose_call) != 0)
1074 			return -EISCONN;
1075 		if (srose.srose_ndigis != rose->dest_ndigis)
1076 			return -EISCONN;
1077 		if (srose.srose_ndigis == rose->dest_ndigis) {
1078 			for (n = 0 ; n < srose.srose_ndigis ; n++)
1079 				if (ax25cmp(&rose->dest_digis[n],
1080 					    &srose.srose_digis[n]))
1081 					return -EISCONN;
1082 		}
1083 		if (srose.srose_family != AF_ROSE)
1084 			return -EINVAL;
1085 	} else {
1086 		if (sk->sk_state != TCP_ESTABLISHED)
1087 			return -ENOTCONN;
1088 
1089 		srose.srose_family = AF_ROSE;
1090 		srose.srose_addr   = rose->dest_addr;
1091 		srose.srose_call   = rose->dest_call;
1092 		srose.srose_ndigis = rose->dest_ndigis;
1093 		for (n = 0 ; n < rose->dest_ndigis ; n++)
1094 			srose.srose_digis[n] = rose->dest_digis[n];
1095 	}
1096 
1097 	SOCK_DEBUG(sk, "ROSE: sendto: Addresses built.\n");
1098 
1099 	/* Build a packet */
1100 	SOCK_DEBUG(sk, "ROSE: sendto: building packet.\n");
1101 	size = len + AX25_BPQ_HEADER_LEN + AX25_MAX_HEADER_LEN + ROSE_MIN_LEN;
1102 
1103 	if ((skb = sock_alloc_send_skb(sk, size, msg->msg_flags & MSG_DONTWAIT, &err)) == NULL)
1104 		return err;
1105 
1106 	skb_reserve(skb, AX25_BPQ_HEADER_LEN + AX25_MAX_HEADER_LEN + ROSE_MIN_LEN);
1107 
1108 	/*
1109 	 *	Put the data on the end
1110 	 */
1111 	SOCK_DEBUG(sk, "ROSE: Appending user data\n");
1112 
1113 	skb_reset_transport_header(skb);
1114 	skb_put(skb, len);
1115 
1116 	err = memcpy_fromiovec(skb_transport_header(skb), msg->msg_iov, len);
1117 	if (err) {
1118 		kfree_skb(skb);
1119 		return err;
1120 	}
1121 
1122 	/*
1123 	 *	If the Q BIT Include socket option is in force, the first
1124 	 *	byte of the user data is the logical value of the Q Bit.
1125 	 */
1126 	if (rose->qbitincl) {
1127 		qbit = skb->data[0];
1128 		skb_pull(skb, 1);
1129 	}
1130 
1131 	/*
1132 	 *	Push down the ROSE header
1133 	 */
1134 	asmptr = skb_push(skb, ROSE_MIN_LEN);
1135 
1136 	SOCK_DEBUG(sk, "ROSE: Building Network Header.\n");
1137 
1138 	/* Build a ROSE Network header */
1139 	asmptr[0] = ((rose->lci >> 8) & 0x0F) | ROSE_GFI;
1140 	asmptr[1] = (rose->lci >> 0) & 0xFF;
1141 	asmptr[2] = ROSE_DATA;
1142 
1143 	if (qbit)
1144 		asmptr[0] |= ROSE_Q_BIT;
1145 
1146 	SOCK_DEBUG(sk, "ROSE: Built header.\n");
1147 
1148 	SOCK_DEBUG(sk, "ROSE: Transmitting buffer\n");
1149 
1150 	if (sk->sk_state != TCP_ESTABLISHED) {
1151 		kfree_skb(skb);
1152 		return -ENOTCONN;
1153 	}
1154 
1155 #ifdef M_BIT
1156 #define ROSE_PACLEN (256-ROSE_MIN_LEN)
1157 	if (skb->len - ROSE_MIN_LEN > ROSE_PACLEN) {
1158 		unsigned char header[ROSE_MIN_LEN];
1159 		struct sk_buff *skbn;
1160 		int frontlen;
1161 		int lg;
1162 
1163 		/* Save a copy of the Header */
1164 		skb_copy_from_linear_data(skb, header, ROSE_MIN_LEN);
1165 		skb_pull(skb, ROSE_MIN_LEN);
1166 
1167 		frontlen = skb_headroom(skb);
1168 
1169 		while (skb->len > 0) {
1170 			if ((skbn = sock_alloc_send_skb(sk, frontlen + ROSE_PACLEN, 0, &err)) == NULL) {
1171 				kfree_skb(skb);
1172 				return err;
1173 			}
1174 
1175 			skbn->sk   = sk;
1176 			skbn->free = 1;
1177 			skbn->arp  = 1;
1178 
1179 			skb_reserve(skbn, frontlen);
1180 
1181 			lg = (ROSE_PACLEN > skb->len) ? skb->len : ROSE_PACLEN;
1182 
1183 			/* Copy the user data */
1184 			skb_copy_from_linear_data(skb, skb_put(skbn, lg), lg);
1185 			skb_pull(skb, lg);
1186 
1187 			/* Duplicate the Header */
1188 			skb_push(skbn, ROSE_MIN_LEN);
1189 			skb_copy_to_linear_data(skbn, header, ROSE_MIN_LEN);
1190 
1191 			if (skb->len > 0)
1192 				skbn->data[2] |= M_BIT;
1193 
1194 			skb_queue_tail(&sk->sk_write_queue, skbn); /* Throw it on the queue */
1195 		}
1196 
1197 		skb->free = 1;
1198 		kfree_skb(skb);
1199 	} else {
1200 		skb_queue_tail(&sk->sk_write_queue, skb);		/* Throw it on the queue */
1201 	}
1202 #else
1203 	skb_queue_tail(&sk->sk_write_queue, skb);	/* Shove it onto the queue */
1204 #endif
1205 
1206 	rose_kick(sk);
1207 
1208 	return len;
1209 }
1210 
1211 
1212 static int rose_recvmsg(struct kiocb *iocb, struct socket *sock,
1213 			struct msghdr *msg, size_t size, int flags)
1214 {
1215 	struct sock *sk = sock->sk;
1216 	struct rose_sock *rose = rose_sk(sk);
1217 	struct sockaddr_rose *srose = (struct sockaddr_rose *)msg->msg_name;
1218 	size_t copied;
1219 	unsigned char *asmptr;
1220 	struct sk_buff *skb;
1221 	int n, er, qbit;
1222 
1223 	/*
1224 	 * This works for seqpacket too. The receiver has ordered the queue for
1225 	 * us! We do one quick check first though
1226 	 */
1227 	if (sk->sk_state != TCP_ESTABLISHED)
1228 		return -ENOTCONN;
1229 
1230 	/* Now we can treat all alike */
1231 	if ((skb = skb_recv_datagram(sk, flags & ~MSG_DONTWAIT, flags & MSG_DONTWAIT, &er)) == NULL)
1232 		return er;
1233 
1234 	qbit = (skb->data[0] & ROSE_Q_BIT) == ROSE_Q_BIT;
1235 
1236 	skb_pull(skb, ROSE_MIN_LEN);
1237 
1238 	if (rose->qbitincl) {
1239 		asmptr  = skb_push(skb, 1);
1240 		*asmptr = qbit;
1241 	}
1242 
1243 	skb_reset_transport_header(skb);
1244 	copied     = skb->len;
1245 
1246 	if (copied > size) {
1247 		copied = size;
1248 		msg->msg_flags |= MSG_TRUNC;
1249 	}
1250 
1251 	skb_copy_datagram_iovec(skb, 0, msg->msg_iov, copied);
1252 
1253 	if (srose != NULL) {
1254 		srose->srose_family = AF_ROSE;
1255 		srose->srose_addr   = rose->dest_addr;
1256 		srose->srose_call   = rose->dest_call;
1257 		srose->srose_ndigis = rose->dest_ndigis;
1258 		if (msg->msg_namelen >= sizeof(struct full_sockaddr_rose)) {
1259 			struct full_sockaddr_rose *full_srose = (struct full_sockaddr_rose *)msg->msg_name;
1260 			for (n = 0 ; n < rose->dest_ndigis ; n++)
1261 				full_srose->srose_digis[n] = rose->dest_digis[n];
1262 			msg->msg_namelen = sizeof(struct full_sockaddr_rose);
1263 		} else {
1264 			if (rose->dest_ndigis >= 1) {
1265 				srose->srose_ndigis = 1;
1266 				srose->srose_digi = rose->dest_digis[0];
1267 			}
1268 			msg->msg_namelen = sizeof(struct sockaddr_rose);
1269 		}
1270 	}
1271 
1272 	skb_free_datagram(sk, skb);
1273 
1274 	return copied;
1275 }
1276 
1277 
1278 static int rose_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg)
1279 {
1280 	struct sock *sk = sock->sk;
1281 	struct rose_sock *rose = rose_sk(sk);
1282 	void __user *argp = (void __user *)arg;
1283 
1284 	switch (cmd) {
1285 	case TIOCOUTQ: {
1286 		long amount;
1287 		amount = sk->sk_sndbuf - atomic_read(&sk->sk_wmem_alloc);
1288 		if (amount < 0)
1289 			amount = 0;
1290 		return put_user(amount, (unsigned int __user *) argp);
1291 	}
1292 
1293 	case TIOCINQ: {
1294 		struct sk_buff *skb;
1295 		long amount = 0L;
1296 		/* These two are safe on a single CPU system as only user tasks fiddle here */
1297 		if ((skb = skb_peek(&sk->sk_receive_queue)) != NULL)
1298 			amount = skb->len;
1299 		return put_user(amount, (unsigned int __user *) argp);
1300 	}
1301 
1302 	case SIOCGSTAMP:
1303 		return sock_get_timestamp(sk, (struct timeval __user *) argp);
1304 
1305 	case SIOCGSTAMPNS:
1306 		return sock_get_timestampns(sk, (struct timespec __user *) argp);
1307 
1308 	case SIOCGIFADDR:
1309 	case SIOCSIFADDR:
1310 	case SIOCGIFDSTADDR:
1311 	case SIOCSIFDSTADDR:
1312 	case SIOCGIFBRDADDR:
1313 	case SIOCSIFBRDADDR:
1314 	case SIOCGIFNETMASK:
1315 	case SIOCSIFNETMASK:
1316 	case SIOCGIFMETRIC:
1317 	case SIOCSIFMETRIC:
1318 		return -EINVAL;
1319 
1320 	case SIOCADDRT:
1321 	case SIOCDELRT:
1322 	case SIOCRSCLRRT:
1323 		if (!capable(CAP_NET_ADMIN))
1324 			return -EPERM;
1325 		return rose_rt_ioctl(cmd, argp);
1326 
1327 	case SIOCRSGCAUSE: {
1328 		struct rose_cause_struct rose_cause;
1329 		rose_cause.cause      = rose->cause;
1330 		rose_cause.diagnostic = rose->diagnostic;
1331 		return copy_to_user(argp, &rose_cause, sizeof(struct rose_cause_struct)) ? -EFAULT : 0;
1332 	}
1333 
1334 	case SIOCRSSCAUSE: {
1335 		struct rose_cause_struct rose_cause;
1336 		if (copy_from_user(&rose_cause, argp, sizeof(struct rose_cause_struct)))
1337 			return -EFAULT;
1338 		rose->cause      = rose_cause.cause;
1339 		rose->diagnostic = rose_cause.diagnostic;
1340 		return 0;
1341 	}
1342 
1343 	case SIOCRSSL2CALL:
1344 		if (!capable(CAP_NET_ADMIN)) return -EPERM;
1345 		if (ax25cmp(&rose_callsign, &null_ax25_address) != 0)
1346 			ax25_listen_release(&rose_callsign, NULL);
1347 		if (copy_from_user(&rose_callsign, argp, sizeof(ax25_address)))
1348 			return -EFAULT;
1349 		if (ax25cmp(&rose_callsign, &null_ax25_address) != 0)
1350 			return ax25_listen_register(&rose_callsign, NULL);
1351 
1352 		return 0;
1353 
1354 	case SIOCRSGL2CALL:
1355 		return copy_to_user(argp, &rose_callsign, sizeof(ax25_address)) ? -EFAULT : 0;
1356 
1357 	case SIOCRSACCEPT:
1358 		if (rose->state == ROSE_STATE_5) {
1359 			rose_write_internal(sk, ROSE_CALL_ACCEPTED);
1360 			rose_start_idletimer(sk);
1361 			rose->condition = 0x00;
1362 			rose->vs        = 0;
1363 			rose->va        = 0;
1364 			rose->vr        = 0;
1365 			rose->vl        = 0;
1366 			rose->state     = ROSE_STATE_3;
1367 		}
1368 		return 0;
1369 
1370 	default:
1371 		return -ENOIOCTLCMD;
1372 	}
1373 
1374 	return 0;
1375 }
1376 
1377 #ifdef CONFIG_PROC_FS
1378 static void *rose_info_start(struct seq_file *seq, loff_t *pos)
1379 {
1380 	int i;
1381 	struct sock *s;
1382 	struct hlist_node *node;
1383 
1384 	spin_lock_bh(&rose_list_lock);
1385 	if (*pos == 0)
1386 		return SEQ_START_TOKEN;
1387 
1388 	i = 1;
1389 	sk_for_each(s, node, &rose_list) {
1390 		if (i == *pos)
1391 			return s;
1392 		++i;
1393 	}
1394 	return NULL;
1395 }
1396 
1397 static void *rose_info_next(struct seq_file *seq, void *v, loff_t *pos)
1398 {
1399 	++*pos;
1400 
1401 	return (v == SEQ_START_TOKEN) ? sk_head(&rose_list)
1402 		: sk_next((struct sock *)v);
1403 }
1404 
1405 static void rose_info_stop(struct seq_file *seq, void *v)
1406 {
1407 	spin_unlock_bh(&rose_list_lock);
1408 }
1409 
1410 static int rose_info_show(struct seq_file *seq, void *v)
1411 {
1412 	char buf[11];
1413 
1414 	if (v == SEQ_START_TOKEN)
1415 		seq_puts(seq,
1416 			 "dest_addr  dest_call src_addr   src_call  dev   lci neigh st vs vr va   t  t1  t2  t3  hb    idle Snd-Q Rcv-Q inode\n");
1417 
1418 	else {
1419 		struct sock *s = v;
1420 		struct rose_sock *rose = rose_sk(s);
1421 		const char *devname, *callsign;
1422 		const struct net_device *dev = rose->device;
1423 
1424 		if (!dev)
1425 			devname = "???";
1426 		else
1427 			devname = dev->name;
1428 
1429 		seq_printf(seq, "%-10s %-9s ",
1430 			rose2asc(&rose->dest_addr),
1431 			ax2asc(buf, &rose->dest_call));
1432 
1433 		if (ax25cmp(&rose->source_call, &null_ax25_address) == 0)
1434 			callsign = "??????-?";
1435 		else
1436 			callsign = ax2asc(buf, &rose->source_call);
1437 
1438 		seq_printf(seq,
1439 			   "%-10s %-9s %-5s %3.3X %05d  %d  %d  %d  %d %3lu %3lu %3lu %3lu %3lu %3lu/%03lu %5d %5d %ld\n",
1440 			rose2asc(&rose->source_addr),
1441 			callsign,
1442 			devname,
1443 			rose->lci & 0x0FFF,
1444 			(rose->neighbour) ? rose->neighbour->number : 0,
1445 			rose->state,
1446 			rose->vs,
1447 			rose->vr,
1448 			rose->va,
1449 			ax25_display_timer(&rose->timer) / HZ,
1450 			rose->t1 / HZ,
1451 			rose->t2 / HZ,
1452 			rose->t3 / HZ,
1453 			rose->hb / HZ,
1454 			ax25_display_timer(&rose->idletimer) / (60 * HZ),
1455 			rose->idle / (60 * HZ),
1456 			atomic_read(&s->sk_wmem_alloc),
1457 			atomic_read(&s->sk_rmem_alloc),
1458 			s->sk_socket ? SOCK_INODE(s->sk_socket)->i_ino : 0L);
1459 	}
1460 
1461 	return 0;
1462 }
1463 
1464 static const struct seq_operations rose_info_seqops = {
1465 	.start = rose_info_start,
1466 	.next = rose_info_next,
1467 	.stop = rose_info_stop,
1468 	.show = rose_info_show,
1469 };
1470 
1471 static int rose_info_open(struct inode *inode, struct file *file)
1472 {
1473 	return seq_open(file, &rose_info_seqops);
1474 }
1475 
1476 static const struct file_operations rose_info_fops = {
1477 	.owner = THIS_MODULE,
1478 	.open = rose_info_open,
1479 	.read = seq_read,
1480 	.llseek = seq_lseek,
1481 	.release = seq_release,
1482 };
1483 #endif	/* CONFIG_PROC_FS */
1484 
1485 static struct net_proto_family rose_family_ops = {
1486 	.family		=	PF_ROSE,
1487 	.create		=	rose_create,
1488 	.owner		=	THIS_MODULE,
1489 };
1490 
1491 static struct proto_ops rose_proto_ops = {
1492 	.family		=	PF_ROSE,
1493 	.owner		=	THIS_MODULE,
1494 	.release	=	rose_release,
1495 	.bind		=	rose_bind,
1496 	.connect	=	rose_connect,
1497 	.socketpair	=	sock_no_socketpair,
1498 	.accept		=	rose_accept,
1499 	.getname	=	rose_getname,
1500 	.poll		=	datagram_poll,
1501 	.ioctl		=	rose_ioctl,
1502 	.listen		=	rose_listen,
1503 	.shutdown	=	sock_no_shutdown,
1504 	.setsockopt	=	rose_setsockopt,
1505 	.getsockopt	=	rose_getsockopt,
1506 	.sendmsg	=	rose_sendmsg,
1507 	.recvmsg	=	rose_recvmsg,
1508 	.mmap		=	sock_no_mmap,
1509 	.sendpage	=	sock_no_sendpage,
1510 };
1511 
1512 static struct notifier_block rose_dev_notifier = {
1513 	.notifier_call	=	rose_device_event,
1514 };
1515 
1516 static struct net_device **dev_rose;
1517 
1518 static struct ax25_protocol rose_pid = {
1519 	.pid	= AX25_P_ROSE,
1520 	.func	= rose_route_frame
1521 };
1522 
1523 static struct ax25_linkfail rose_linkfail_notifier = {
1524 	.func	= rose_link_failed
1525 };
1526 
1527 static int __init rose_proto_init(void)
1528 {
1529 	int i;
1530 	int rc;
1531 
1532 	if (rose_ndevs > 0x7FFFFFFF/sizeof(struct net_device *)) {
1533 		printk(KERN_ERR "ROSE: rose_proto_init - rose_ndevs parameter to large\n");
1534 		rc = -EINVAL;
1535 		goto out;
1536 	}
1537 
1538 	rc = proto_register(&rose_proto, 0);
1539 	if (rc != 0)
1540 		goto out;
1541 
1542 	rose_callsign = null_ax25_address;
1543 
1544 	dev_rose = kzalloc(rose_ndevs * sizeof(struct net_device *), GFP_KERNEL);
1545 	if (dev_rose == NULL) {
1546 		printk(KERN_ERR "ROSE: rose_proto_init - unable to allocate device structure\n");
1547 		rc = -ENOMEM;
1548 		goto out_proto_unregister;
1549 	}
1550 
1551 	for (i = 0; i < rose_ndevs; i++) {
1552 		struct net_device *dev;
1553 		char name[IFNAMSIZ];
1554 
1555 		sprintf(name, "rose%d", i);
1556 		dev = alloc_netdev(sizeof(struct net_device_stats),
1557 				   name, rose_setup);
1558 		if (!dev) {
1559 			printk(KERN_ERR "ROSE: rose_proto_init - unable to allocate memory\n");
1560 			rc = -ENOMEM;
1561 			goto fail;
1562 		}
1563 		rc = register_netdev(dev);
1564 		if (rc) {
1565 			printk(KERN_ERR "ROSE: netdevice registration failed\n");
1566 			free_netdev(dev);
1567 			goto fail;
1568 		}
1569 		lockdep_set_class(&dev->_xmit_lock, &rose_netdev_xmit_lock_key);
1570 		dev_rose[i] = dev;
1571 	}
1572 
1573 	sock_register(&rose_family_ops);
1574 	register_netdevice_notifier(&rose_dev_notifier);
1575 
1576 	ax25_register_pid(&rose_pid);
1577 	ax25_linkfail_register(&rose_linkfail_notifier);
1578 
1579 #ifdef CONFIG_SYSCTL
1580 	rose_register_sysctl();
1581 #endif
1582 	rose_loopback_init();
1583 
1584 	rose_add_loopback_neigh();
1585 
1586 	proc_net_fops_create(&init_net, "rose", S_IRUGO, &rose_info_fops);
1587 	proc_net_fops_create(&init_net, "rose_neigh", S_IRUGO, &rose_neigh_fops);
1588 	proc_net_fops_create(&init_net, "rose_nodes", S_IRUGO, &rose_nodes_fops);
1589 	proc_net_fops_create(&init_net, "rose_routes", S_IRUGO, &rose_routes_fops);
1590 out:
1591 	return rc;
1592 fail:
1593 	while (--i >= 0) {
1594 		unregister_netdev(dev_rose[i]);
1595 		free_netdev(dev_rose[i]);
1596 	}
1597 	kfree(dev_rose);
1598 out_proto_unregister:
1599 	proto_unregister(&rose_proto);
1600 	goto out;
1601 }
1602 module_init(rose_proto_init);
1603 
1604 module_param(rose_ndevs, int, 0);
1605 MODULE_PARM_DESC(rose_ndevs, "number of ROSE devices");
1606 
1607 MODULE_AUTHOR("Jonathan Naylor G4KLX <g4klx@g4klx.demon.co.uk>");
1608 MODULE_DESCRIPTION("The amateur radio ROSE network layer protocol");
1609 MODULE_LICENSE("GPL");
1610 MODULE_ALIAS_NETPROTO(PF_ROSE);
1611 
1612 static void __exit rose_exit(void)
1613 {
1614 	int i;
1615 
1616 	proc_net_remove(&init_net, "rose");
1617 	proc_net_remove(&init_net, "rose_neigh");
1618 	proc_net_remove(&init_net, "rose_nodes");
1619 	proc_net_remove(&init_net, "rose_routes");
1620 	rose_loopback_clear();
1621 
1622 	rose_rt_free();
1623 
1624 	ax25_protocol_release(AX25_P_ROSE);
1625 	ax25_linkfail_release(&rose_linkfail_notifier);
1626 
1627 	if (ax25cmp(&rose_callsign, &null_ax25_address) != 0)
1628 		ax25_listen_release(&rose_callsign, NULL);
1629 
1630 #ifdef CONFIG_SYSCTL
1631 	rose_unregister_sysctl();
1632 #endif
1633 	unregister_netdevice_notifier(&rose_dev_notifier);
1634 
1635 	sock_unregister(PF_ROSE);
1636 
1637 	for (i = 0; i < rose_ndevs; i++) {
1638 		struct net_device *dev = dev_rose[i];
1639 
1640 		if (dev) {
1641 			unregister_netdev(dev);
1642 			free_netdev(dev);
1643 		}
1644 	}
1645 
1646 	kfree(dev_rose);
1647 	proto_unregister(&rose_proto);
1648 }
1649 
1650 module_exit(rose_exit);
1651