Merge branch 'master' of git://git.kernel.org/pub/scm/linux/kernel/git/rusty/linux...
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / net / decnet / af_decnet.c
1
2 /*
3 * DECnet An implementation of the DECnet protocol suite for the LINUX
4 * operating system. DECnet is implemented using the BSD Socket
5 * interface as the means of communication with the user level.
6 *
7 * DECnet Socket Layer Interface
8 *
9 * Authors: Eduardo Marcelo Serrat <emserrat@geocities.com>
10 * Patrick Caulfield <patrick@pandh.demon.co.uk>
11 *
12 * Changes:
13 * Steve Whitehouse: Copied from Eduardo Serrat and Patrick Caulfield's
14 * version of the code. Original copyright preserved
15 * below.
16 * Steve Whitehouse: Some bug fixes, cleaning up some code to make it
17 * compatible with my routing layer.
18 * Steve Whitehouse: Merging changes from Eduardo Serrat and Patrick
19 * Caulfield.
20 * Steve Whitehouse: Further bug fixes, checking module code still works
21 * with new routing layer.
22 * Steve Whitehouse: Additional set/get_sockopt() calls.
23 * Steve Whitehouse: Fixed TIOCINQ ioctl to be same as Eduardo's new
24 * code.
25 * Steve Whitehouse: recvmsg() changed to try and behave in a POSIX like
26 * way. Didn't manage it entirely, but its better.
27 * Steve Whitehouse: ditto for sendmsg().
28 * Steve Whitehouse: A selection of bug fixes to various things.
29 * Steve Whitehouse: Added TIOCOUTQ ioctl.
30 * Steve Whitehouse: Fixes to username2sockaddr & sockaddr2username.
31 * Steve Whitehouse: Fixes to connect() error returns.
32 * Patrick Caulfield: Fixes to delayed acceptance logic.
33 * David S. Miller: New socket locking
34 * Steve Whitehouse: Socket list hashing/locking
35 * Arnaldo C. Melo: use capable, not suser
36 * Steve Whitehouse: Removed unused code. Fix to use sk->allocation
37 * when required.
38 * Patrick Caulfield: /proc/net/decnet now has object name/number
39 * Steve Whitehouse: Fixed local port allocation, hashed sk list
40 * Matthew Wilcox: Fixes for dn_ioctl()
41 * Steve Whitehouse: New connect/accept logic to allow timeouts and
42 * prepare for sendpage etc.
43 */
44
45
46 /******************************************************************************
47 (c) 1995-1998 E.M. Serrat emserrat@geocities.com
48
49 This program is free software; you can redistribute it and/or modify
50 it under the terms of the GNU General Public License as published by
51 the Free Software Foundation; either version 2 of the License, or
52 any later version.
53
54 This program is distributed in the hope that it will be useful,
55 but WITHOUT ANY WARRANTY; without even the implied warranty of
56 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
57 GNU General Public License for more details.
58
59 HISTORY:
60
61 Version Kernel Date Author/Comments
62 ------- ------ ---- ---------------
63 Version 0.0.1 2.0.30 01-dic-97 Eduardo Marcelo Serrat
64 (emserrat@geocities.com)
65
66 First Development of DECnet Socket La-
67 yer for Linux. Only supports outgoing
68 connections.
69
70 Version 0.0.2 2.1.105 20-jun-98 Patrick J. Caulfield
71 (patrick@pandh.demon.co.uk)
72
73 Port to new kernel development version.
74
75 Version 0.0.3 2.1.106 25-jun-98 Eduardo Marcelo Serrat
76 (emserrat@geocities.com)
77 _
78 Added support for incoming connections
79 so we can start developing server apps
80 on Linux.
81 -
82 Module Support
83 Version 0.0.4 2.1.109 21-jul-98 Eduardo Marcelo Serrat
84 (emserrat@geocities.com)
85 _
86 Added support for X11R6.4. Now we can
87 use DECnet transport for X on Linux!!!
88 -
89 Version 0.0.5 2.1.110 01-aug-98 Eduardo Marcelo Serrat
90 (emserrat@geocities.com)
91 Removed bugs on flow control
92 Removed bugs on incoming accessdata
93 order
94 -
95 Version 0.0.6 2.1.110 07-aug-98 Eduardo Marcelo Serrat
96 dn_recvmsg fixes
97
98 Patrick J. Caulfield
99 dn_bind fixes
100 *******************************************************************************/
101
102 #include <linux/module.h>
103 #include <linux/errno.h>
104 #include <linux/types.h>
105 #include <linux/slab.h>
106 #include <linux/socket.h>
107 #include <linux/in.h>
108 #include <linux/kernel.h>
109 #include <linux/sched.h>
110 #include <linux/timer.h>
111 #include <linux/string.h>
112 #include <linux/sockios.h>
113 #include <linux/net.h>
114 #include <linux/netdevice.h>
115 #include <linux/inet.h>
116 #include <linux/route.h>
117 #include <linux/netfilter.h>
118 #include <linux/seq_file.h>
119 #include <net/sock.h>
120 #include <net/tcp_states.h>
121 #include <net/flow.h>
122 #include <asm/system.h>
123 #include <asm/ioctls.h>
124 #include <linux/capability.h>
125 #include <linux/mm.h>
126 #include <linux/interrupt.h>
127 #include <linux/proc_fs.h>
128 #include <linux/stat.h>
129 #include <linux/init.h>
130 #include <linux/poll.h>
131 #include <net/net_namespace.h>
132 #include <net/neighbour.h>
133 #include <net/dst.h>
134 #include <net/fib_rules.h>
135 #include <net/dn.h>
136 #include <net/dn_nsp.h>
137 #include <net/dn_dev.h>
138 #include <net/dn_route.h>
139 #include <net/dn_fib.h>
140 #include <net/dn_neigh.h>
141
142 struct dn_sock {
143 struct sock sk;
144 struct dn_scp scp;
145 };
146
147 static void dn_keepalive(struct sock *sk);
148
149 #define DN_SK_HASH_SHIFT 8
150 #define DN_SK_HASH_SIZE (1 << DN_SK_HASH_SHIFT)
151 #define DN_SK_HASH_MASK (DN_SK_HASH_SIZE - 1)
152
153
154 static const struct proto_ops dn_proto_ops;
155 static DEFINE_RWLOCK(dn_hash_lock);
156 static struct hlist_head dn_sk_hash[DN_SK_HASH_SIZE];
157 static struct hlist_head dn_wild_sk;
158 static atomic_t decnet_memory_allocated;
159
160 static int __dn_setsockopt(struct socket *sock, int level, int optname, char __user *optval, int optlen, int flags);
161 static int __dn_getsockopt(struct socket *sock, int level, int optname, char __user *optval, int __user *optlen, int flags);
162
163 static struct hlist_head *dn_find_list(struct sock *sk)
164 {
165 struct dn_scp *scp = DN_SK(sk);
166
167 if (scp->addr.sdn_flags & SDF_WILD)
168 return hlist_empty(&dn_wild_sk) ? &dn_wild_sk : NULL;
169
170 return &dn_sk_hash[le16_to_cpu(scp->addrloc) & DN_SK_HASH_MASK];
171 }
172
173 /*
174 * Valid ports are those greater than zero and not already in use.
175 */
176 static int check_port(__le16 port)
177 {
178 struct sock *sk;
179 struct hlist_node *node;
180
181 if (port == 0)
182 return -1;
183
184 sk_for_each(sk, node, &dn_sk_hash[le16_to_cpu(port) & DN_SK_HASH_MASK]) {
185 struct dn_scp *scp = DN_SK(sk);
186 if (scp->addrloc == port)
187 return -1;
188 }
189 return 0;
190 }
191
192 static unsigned short port_alloc(struct sock *sk)
193 {
194 struct dn_scp *scp = DN_SK(sk);
195 static unsigned short port = 0x2000;
196 unsigned short i_port = port;
197
198 while(check_port(cpu_to_le16(++port)) != 0) {
199 if (port == i_port)
200 return 0;
201 }
202
203 scp->addrloc = cpu_to_le16(port);
204
205 return 1;
206 }
207
208 /*
209 * Since this is only ever called from user
210 * level, we don't need a write_lock() version
211 * of this.
212 */
213 static int dn_hash_sock(struct sock *sk)
214 {
215 struct dn_scp *scp = DN_SK(sk);
216 struct hlist_head *list;
217 int rv = -EUSERS;
218
219 BUG_ON(sk_hashed(sk));
220
221 write_lock_bh(&dn_hash_lock);
222
223 if (!scp->addrloc && !port_alloc(sk))
224 goto out;
225
226 rv = -EADDRINUSE;
227 if ((list = dn_find_list(sk)) == NULL)
228 goto out;
229
230 sk_add_node(sk, list);
231 rv = 0;
232 out:
233 write_unlock_bh(&dn_hash_lock);
234 return rv;
235 }
236
237 static void dn_unhash_sock(struct sock *sk)
238 {
239 write_lock(&dn_hash_lock);
240 sk_del_node_init(sk);
241 write_unlock(&dn_hash_lock);
242 }
243
244 static void dn_unhash_sock_bh(struct sock *sk)
245 {
246 write_lock_bh(&dn_hash_lock);
247 sk_del_node_init(sk);
248 write_unlock_bh(&dn_hash_lock);
249 }
250
251 static struct hlist_head *listen_hash(struct sockaddr_dn *addr)
252 {
253 int i;
254 unsigned hash = addr->sdn_objnum;
255
256 if (hash == 0) {
257 hash = addr->sdn_objnamel;
258 for(i = 0; i < le16_to_cpu(addr->sdn_objnamel); i++) {
259 hash ^= addr->sdn_objname[i];
260 hash ^= (hash << 3);
261 }
262 }
263
264 return &dn_sk_hash[hash & DN_SK_HASH_MASK];
265 }
266
267 /*
268 * Called to transform a socket from bound (i.e. with a local address)
269 * into a listening socket (doesn't need a local port number) and rehashes
270 * based upon the object name/number.
271 */
272 static void dn_rehash_sock(struct sock *sk)
273 {
274 struct hlist_head *list;
275 struct dn_scp *scp = DN_SK(sk);
276
277 if (scp->addr.sdn_flags & SDF_WILD)
278 return;
279
280 write_lock_bh(&dn_hash_lock);
281 sk_del_node_init(sk);
282 DN_SK(sk)->addrloc = 0;
283 list = listen_hash(&DN_SK(sk)->addr);
284 sk_add_node(sk, list);
285 write_unlock_bh(&dn_hash_lock);
286 }
287
288 int dn_sockaddr2username(struct sockaddr_dn *sdn, unsigned char *buf, unsigned char type)
289 {
290 int len = 2;
291
292 *buf++ = type;
293
294 switch(type) {
295 case 0:
296 *buf++ = sdn->sdn_objnum;
297 break;
298 case 1:
299 *buf++ = 0;
300 *buf++ = le16_to_cpu(sdn->sdn_objnamel);
301 memcpy(buf, sdn->sdn_objname, le16_to_cpu(sdn->sdn_objnamel));
302 len = 3 + le16_to_cpu(sdn->sdn_objnamel);
303 break;
304 case 2:
305 memset(buf, 0, 5);
306 buf += 5;
307 *buf++ = le16_to_cpu(sdn->sdn_objnamel);
308 memcpy(buf, sdn->sdn_objname, le16_to_cpu(sdn->sdn_objnamel));
309 len = 7 + le16_to_cpu(sdn->sdn_objnamel);
310 break;
311 }
312
313 return len;
314 }
315
316 /*
317 * On reception of usernames, we handle types 1 and 0 for destination
318 * addresses only. Types 2 and 4 are used for source addresses, but the
319 * UIC, GIC are ignored and they are both treated the same way. Type 3
320 * is never used as I've no idea what its purpose might be or what its
321 * format is.
322 */
323 int dn_username2sockaddr(unsigned char *data, int len, struct sockaddr_dn *sdn, unsigned char *fmt)
324 {
325 unsigned char type;
326 int size = len;
327 int namel = 12;
328
329 sdn->sdn_objnum = 0;
330 sdn->sdn_objnamel = cpu_to_le16(0);
331 memset(sdn->sdn_objname, 0, DN_MAXOBJL);
332
333 if (len < 2)
334 return -1;
335
336 len -= 2;
337 *fmt = *data++;
338 type = *data++;
339
340 switch(*fmt) {
341 case 0:
342 sdn->sdn_objnum = type;
343 return 2;
344 case 1:
345 namel = 16;
346 break;
347 case 2:
348 len -= 4;
349 data += 4;
350 break;
351 case 4:
352 len -= 8;
353 data += 8;
354 break;
355 default:
356 return -1;
357 }
358
359 len -= 1;
360
361 if (len < 0)
362 return -1;
363
364 sdn->sdn_objnamel = cpu_to_le16(*data++);
365 len -= le16_to_cpu(sdn->sdn_objnamel);
366
367 if ((len < 0) || (le16_to_cpu(sdn->sdn_objnamel) > namel))
368 return -1;
369
370 memcpy(sdn->sdn_objname, data, le16_to_cpu(sdn->sdn_objnamel));
371
372 return size - len;
373 }
374
375 struct sock *dn_sklist_find_listener(struct sockaddr_dn *addr)
376 {
377 struct hlist_head *list = listen_hash(addr);
378 struct hlist_node *node;
379 struct sock *sk;
380
381 read_lock(&dn_hash_lock);
382 sk_for_each(sk, node, list) {
383 struct dn_scp *scp = DN_SK(sk);
384 if (sk->sk_state != TCP_LISTEN)
385 continue;
386 if (scp->addr.sdn_objnum) {
387 if (scp->addr.sdn_objnum != addr->sdn_objnum)
388 continue;
389 } else {
390 if (addr->sdn_objnum)
391 continue;
392 if (scp->addr.sdn_objnamel != addr->sdn_objnamel)
393 continue;
394 if (memcmp(scp->addr.sdn_objname, addr->sdn_objname, le16_to_cpu(addr->sdn_objnamel)) != 0)
395 continue;
396 }
397 sock_hold(sk);
398 read_unlock(&dn_hash_lock);
399 return sk;
400 }
401
402 sk = sk_head(&dn_wild_sk);
403 if (sk) {
404 if (sk->sk_state == TCP_LISTEN)
405 sock_hold(sk);
406 else
407 sk = NULL;
408 }
409
410 read_unlock(&dn_hash_lock);
411 return sk;
412 }
413
414 struct sock *dn_find_by_skb(struct sk_buff *skb)
415 {
416 struct dn_skb_cb *cb = DN_SKB_CB(skb);
417 struct sock *sk;
418 struct hlist_node *node;
419 struct dn_scp *scp;
420
421 read_lock(&dn_hash_lock);
422 sk_for_each(sk, node, &dn_sk_hash[le16_to_cpu(cb->dst_port) & DN_SK_HASH_MASK]) {
423 scp = DN_SK(sk);
424 if (cb->src != dn_saddr2dn(&scp->peer))
425 continue;
426 if (cb->dst_port != scp->addrloc)
427 continue;
428 if (scp->addrrem && (cb->src_port != scp->addrrem))
429 continue;
430 sock_hold(sk);
431 goto found;
432 }
433 sk = NULL;
434 found:
435 read_unlock(&dn_hash_lock);
436 return sk;
437 }
438
439
440
441 static void dn_destruct(struct sock *sk)
442 {
443 struct dn_scp *scp = DN_SK(sk);
444
445 skb_queue_purge(&scp->data_xmit_queue);
446 skb_queue_purge(&scp->other_xmit_queue);
447 skb_queue_purge(&scp->other_receive_queue);
448
449 dst_release(xchg(&sk->sk_dst_cache, NULL));
450 }
451
452 static int dn_memory_pressure;
453
454 static void dn_enter_memory_pressure(struct sock *sk)
455 {
456 if (!dn_memory_pressure) {
457 dn_memory_pressure = 1;
458 }
459 }
460
461 static struct proto dn_proto = {
462 .name = "NSP",
463 .owner = THIS_MODULE,
464 .enter_memory_pressure = dn_enter_memory_pressure,
465 .memory_pressure = &dn_memory_pressure,
466 .memory_allocated = &decnet_memory_allocated,
467 .sysctl_mem = sysctl_decnet_mem,
468 .sysctl_wmem = sysctl_decnet_wmem,
469 .sysctl_rmem = sysctl_decnet_rmem,
470 .max_header = DN_MAX_NSP_DATA_HEADER + 64,
471 .obj_size = sizeof(struct dn_sock),
472 };
473
474 static struct sock *dn_alloc_sock(struct net *net, struct socket *sock, gfp_t gfp)
475 {
476 struct dn_scp *scp;
477 struct sock *sk = sk_alloc(net, PF_DECnet, gfp, &dn_proto);
478
479 if (!sk)
480 goto out;
481
482 if (sock)
483 sock->ops = &dn_proto_ops;
484 sock_init_data(sock, sk);
485
486 sk->sk_backlog_rcv = dn_nsp_backlog_rcv;
487 sk->sk_destruct = dn_destruct;
488 sk->sk_no_check = 1;
489 sk->sk_family = PF_DECnet;
490 sk->sk_protocol = 0;
491 sk->sk_allocation = gfp;
492 sk->sk_sndbuf = sysctl_decnet_wmem[1];
493 sk->sk_rcvbuf = sysctl_decnet_rmem[1];
494
495 /* Initialization of DECnet Session Control Port */
496 scp = DN_SK(sk);
497 scp->state = DN_O; /* Open */
498 scp->numdat = 1; /* Next data seg to tx */
499 scp->numoth = 1; /* Next oth data to tx */
500 scp->ackxmt_dat = 0; /* Last data seg ack'ed */
501 scp->ackxmt_oth = 0; /* Last oth data ack'ed */
502 scp->ackrcv_dat = 0; /* Highest data ack recv*/
503 scp->ackrcv_oth = 0; /* Last oth data ack rec*/
504 scp->flowrem_sw = DN_SEND;
505 scp->flowloc_sw = DN_SEND;
506 scp->flowrem_dat = 0;
507 scp->flowrem_oth = 1;
508 scp->flowloc_dat = 0;
509 scp->flowloc_oth = 1;
510 scp->services_rem = 0;
511 scp->services_loc = 1 | NSP_FC_NONE;
512 scp->info_rem = 0;
513 scp->info_loc = 0x03; /* NSP version 4.1 */
514 scp->segsize_rem = 230 - DN_MAX_NSP_DATA_HEADER; /* Default: Updated by remote segsize */
515 scp->nonagle = 0;
516 scp->multi_ireq = 1;
517 scp->accept_mode = ACC_IMMED;
518 scp->addr.sdn_family = AF_DECnet;
519 scp->peer.sdn_family = AF_DECnet;
520 scp->accessdata.acc_accl = 5;
521 memcpy(scp->accessdata.acc_acc, "LINUX", 5);
522
523 scp->max_window = NSP_MAX_WINDOW;
524 scp->snd_window = NSP_MIN_WINDOW;
525 scp->nsp_srtt = NSP_INITIAL_SRTT;
526 scp->nsp_rttvar = NSP_INITIAL_RTTVAR;
527 scp->nsp_rxtshift = 0;
528
529 skb_queue_head_init(&scp->data_xmit_queue);
530 skb_queue_head_init(&scp->other_xmit_queue);
531 skb_queue_head_init(&scp->other_receive_queue);
532
533 scp->persist = 0;
534 scp->persist_fxn = NULL;
535 scp->keepalive = 10 * HZ;
536 scp->keepalive_fxn = dn_keepalive;
537
538 init_timer(&scp->delack_timer);
539 scp->delack_pending = 0;
540 scp->delack_fxn = dn_nsp_delayed_ack;
541
542 dn_start_slow_timer(sk);
543 out:
544 return sk;
545 }
546
547 /*
548 * Keepalive timer.
549 * FIXME: Should respond to SO_KEEPALIVE etc.
550 */
551 static void dn_keepalive(struct sock *sk)
552 {
553 struct dn_scp *scp = DN_SK(sk);
554
555 /*
556 * By checking the other_data transmit queue is empty
557 * we are double checking that we are not sending too
558 * many of these keepalive frames.
559 */
560 if (skb_queue_empty(&scp->other_xmit_queue))
561 dn_nsp_send_link(sk, DN_NOCHANGE, 0);
562 }
563
564
565 /*
566 * Timer for shutdown/destroyed sockets.
567 * When socket is dead & no packets have been sent for a
568 * certain amount of time, they are removed by this
569 * routine. Also takes care of sending out DI & DC
570 * frames at correct times.
571 */
572 int dn_destroy_timer(struct sock *sk)
573 {
574 struct dn_scp *scp = DN_SK(sk);
575
576 scp->persist = dn_nsp_persist(sk);
577
578 switch(scp->state) {
579 case DN_DI:
580 dn_nsp_send_disc(sk, NSP_DISCINIT, 0, GFP_ATOMIC);
581 if (scp->nsp_rxtshift >= decnet_di_count)
582 scp->state = DN_CN;
583 return 0;
584
585 case DN_DR:
586 dn_nsp_send_disc(sk, NSP_DISCINIT, 0, GFP_ATOMIC);
587 if (scp->nsp_rxtshift >= decnet_dr_count)
588 scp->state = DN_DRC;
589 return 0;
590
591 case DN_DN:
592 if (scp->nsp_rxtshift < decnet_dn_count) {
593 /* printk(KERN_DEBUG "dn_destroy_timer: DN\n"); */
594 dn_nsp_send_disc(sk, NSP_DISCCONF, NSP_REASON_DC, GFP_ATOMIC);
595 return 0;
596 }
597 }
598
599 scp->persist = (HZ * decnet_time_wait);
600
601 if (sk->sk_socket)
602 return 0;
603
604 if ((jiffies - scp->stamp) >= (HZ * decnet_time_wait)) {
605 dn_unhash_sock(sk);
606 sock_put(sk);
607 return 1;
608 }
609
610 return 0;
611 }
612
613 static void dn_destroy_sock(struct sock *sk)
614 {
615 struct dn_scp *scp = DN_SK(sk);
616
617 scp->nsp_rxtshift = 0; /* reset back off */
618
619 if (sk->sk_socket) {
620 if (sk->sk_socket->state != SS_UNCONNECTED)
621 sk->sk_socket->state = SS_DISCONNECTING;
622 }
623
624 sk->sk_state = TCP_CLOSE;
625
626 switch(scp->state) {
627 case DN_DN:
628 dn_nsp_send_disc(sk, NSP_DISCCONF, NSP_REASON_DC,
629 sk->sk_allocation);
630 scp->persist_fxn = dn_destroy_timer;
631 scp->persist = dn_nsp_persist(sk);
632 break;
633 case DN_CR:
634 scp->state = DN_DR;
635 goto disc_reject;
636 case DN_RUN:
637 scp->state = DN_DI;
638 case DN_DI:
639 case DN_DR:
640 disc_reject:
641 dn_nsp_send_disc(sk, NSP_DISCINIT, 0, sk->sk_allocation);
642 case DN_NC:
643 case DN_NR:
644 case DN_RJ:
645 case DN_DIC:
646 case DN_CN:
647 case DN_DRC:
648 case DN_CI:
649 case DN_CD:
650 scp->persist_fxn = dn_destroy_timer;
651 scp->persist = dn_nsp_persist(sk);
652 break;
653 default:
654 printk(KERN_DEBUG "DECnet: dn_destroy_sock passed socket in invalid state\n");
655 case DN_O:
656 dn_stop_slow_timer(sk);
657
658 dn_unhash_sock_bh(sk);
659 sock_put(sk);
660
661 break;
662 }
663 }
664
665 char *dn_addr2asc(__u16 addr, char *buf)
666 {
667 unsigned short node, area;
668
669 node = addr & 0x03ff;
670 area = addr >> 10;
671 sprintf(buf, "%hd.%hd", area, node);
672
673 return buf;
674 }
675
676
677
678 static int dn_create(struct net *net, struct socket *sock, int protocol)
679 {
680 struct sock *sk;
681
682 if (net != &init_net)
683 return -EAFNOSUPPORT;
684
685 switch(sock->type) {
686 case SOCK_SEQPACKET:
687 if (protocol != DNPROTO_NSP)
688 return -EPROTONOSUPPORT;
689 break;
690 case SOCK_STREAM:
691 break;
692 default:
693 return -ESOCKTNOSUPPORT;
694 }
695
696
697 if ((sk = dn_alloc_sock(net, sock, GFP_KERNEL)) == NULL)
698 return -ENOBUFS;
699
700 sk->sk_protocol = protocol;
701
702 return 0;
703 }
704
705
706 static int
707 dn_release(struct socket *sock)
708 {
709 struct sock *sk = sock->sk;
710
711 if (sk) {
712 sock_orphan(sk);
713 sock_hold(sk);
714 lock_sock(sk);
715 dn_destroy_sock(sk);
716 release_sock(sk);
717 sock_put(sk);
718 }
719
720 return 0;
721 }
722
723 static int dn_bind(struct socket *sock, struct sockaddr *uaddr, int addr_len)
724 {
725 struct sock *sk = sock->sk;
726 struct dn_scp *scp = DN_SK(sk);
727 struct sockaddr_dn *saddr = (struct sockaddr_dn *)uaddr;
728 struct net_device *dev, *ldev;
729 int rv;
730
731 if (addr_len != sizeof(struct sockaddr_dn))
732 return -EINVAL;
733
734 if (saddr->sdn_family != AF_DECnet)
735 return -EINVAL;
736
737 if (le16_to_cpu(saddr->sdn_nodeaddrl) && (le16_to_cpu(saddr->sdn_nodeaddrl) != 2))
738 return -EINVAL;
739
740 if (le16_to_cpu(saddr->sdn_objnamel) > DN_MAXOBJL)
741 return -EINVAL;
742
743 if (saddr->sdn_flags & ~SDF_WILD)
744 return -EINVAL;
745
746 if (!capable(CAP_NET_BIND_SERVICE) && (saddr->sdn_objnum ||
747 (saddr->sdn_flags & SDF_WILD)))
748 return -EACCES;
749
750 if (!(saddr->sdn_flags & SDF_WILD)) {
751 if (le16_to_cpu(saddr->sdn_nodeaddrl)) {
752 read_lock(&dev_base_lock);
753 ldev = NULL;
754 for_each_netdev(&init_net, dev) {
755 if (!dev->dn_ptr)
756 continue;
757 if (dn_dev_islocal(dev, dn_saddr2dn(saddr))) {
758 ldev = dev;
759 break;
760 }
761 }
762 read_unlock(&dev_base_lock);
763 if (ldev == NULL)
764 return -EADDRNOTAVAIL;
765 }
766 }
767
768 rv = -EINVAL;
769 lock_sock(sk);
770 if (sock_flag(sk, SOCK_ZAPPED)) {
771 memcpy(&scp->addr, saddr, addr_len);
772 sock_reset_flag(sk, SOCK_ZAPPED);
773
774 rv = dn_hash_sock(sk);
775 if (rv)
776 sock_set_flag(sk, SOCK_ZAPPED);
777 }
778 release_sock(sk);
779
780 return rv;
781 }
782
783
784 static int dn_auto_bind(struct socket *sock)
785 {
786 struct sock *sk = sock->sk;
787 struct dn_scp *scp = DN_SK(sk);
788 int rv;
789
790 sock_reset_flag(sk, SOCK_ZAPPED);
791
792 scp->addr.sdn_flags = 0;
793 scp->addr.sdn_objnum = 0;
794
795 /*
796 * This stuff is to keep compatibility with Eduardo's
797 * patch. I hope I can dispense with it shortly...
798 */
799 if ((scp->accessdata.acc_accl != 0) &&
800 (scp->accessdata.acc_accl <= 12)) {
801
802 scp->addr.sdn_objnamel = cpu_to_le16(scp->accessdata.acc_accl);
803 memcpy(scp->addr.sdn_objname, scp->accessdata.acc_acc, le16_to_cpu(scp->addr.sdn_objnamel));
804
805 scp->accessdata.acc_accl = 0;
806 memset(scp->accessdata.acc_acc, 0, 40);
807 }
808 /* End of compatibility stuff */
809
810 scp->addr.sdn_add.a_len = cpu_to_le16(2);
811 rv = dn_dev_bind_default((__le16 *)scp->addr.sdn_add.a_addr);
812 if (rv == 0) {
813 rv = dn_hash_sock(sk);
814 if (rv)
815 sock_set_flag(sk, SOCK_ZAPPED);
816 }
817
818 return rv;
819 }
820
821 static int dn_confirm_accept(struct sock *sk, long *timeo, gfp_t allocation)
822 {
823 struct dn_scp *scp = DN_SK(sk);
824 DEFINE_WAIT(wait);
825 int err;
826
827 if (scp->state != DN_CR)
828 return -EINVAL;
829
830 scp->state = DN_CC;
831 scp->segsize_loc = dst_metric(__sk_dst_get(sk), RTAX_ADVMSS);
832 dn_send_conn_conf(sk, allocation);
833
834 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
835 for(;;) {
836 release_sock(sk);
837 if (scp->state == DN_CC)
838 *timeo = schedule_timeout(*timeo);
839 lock_sock(sk);
840 err = 0;
841 if (scp->state == DN_RUN)
842 break;
843 err = sock_error(sk);
844 if (err)
845 break;
846 err = sock_intr_errno(*timeo);
847 if (signal_pending(current))
848 break;
849 err = -EAGAIN;
850 if (!*timeo)
851 break;
852 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
853 }
854 finish_wait(sk->sk_sleep, &wait);
855 if (err == 0) {
856 sk->sk_socket->state = SS_CONNECTED;
857 } else if (scp->state != DN_CC) {
858 sk->sk_socket->state = SS_UNCONNECTED;
859 }
860 return err;
861 }
862
863 static int dn_wait_run(struct sock *sk, long *timeo)
864 {
865 struct dn_scp *scp = DN_SK(sk);
866 DEFINE_WAIT(wait);
867 int err = 0;
868
869 if (scp->state == DN_RUN)
870 goto out;
871
872 if (!*timeo)
873 return -EALREADY;
874
875 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
876 for(;;) {
877 release_sock(sk);
878 if (scp->state == DN_CI || scp->state == DN_CC)
879 *timeo = schedule_timeout(*timeo);
880 lock_sock(sk);
881 err = 0;
882 if (scp->state == DN_RUN)
883 break;
884 err = sock_error(sk);
885 if (err)
886 break;
887 err = sock_intr_errno(*timeo);
888 if (signal_pending(current))
889 break;
890 err = -ETIMEDOUT;
891 if (!*timeo)
892 break;
893 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
894 }
895 finish_wait(sk->sk_sleep, &wait);
896 out:
897 if (err == 0) {
898 sk->sk_socket->state = SS_CONNECTED;
899 } else if (scp->state != DN_CI && scp->state != DN_CC) {
900 sk->sk_socket->state = SS_UNCONNECTED;
901 }
902 return err;
903 }
904
905 static int __dn_connect(struct sock *sk, struct sockaddr_dn *addr, int addrlen, long *timeo, int flags)
906 {
907 struct socket *sock = sk->sk_socket;
908 struct dn_scp *scp = DN_SK(sk);
909 int err = -EISCONN;
910 struct flowi fl;
911
912 if (sock->state == SS_CONNECTED)
913 goto out;
914
915 if (sock->state == SS_CONNECTING) {
916 err = 0;
917 if (scp->state == DN_RUN) {
918 sock->state = SS_CONNECTED;
919 goto out;
920 }
921 err = -ECONNREFUSED;
922 if (scp->state != DN_CI && scp->state != DN_CC) {
923 sock->state = SS_UNCONNECTED;
924 goto out;
925 }
926 return dn_wait_run(sk, timeo);
927 }
928
929 err = -EINVAL;
930 if (scp->state != DN_O)
931 goto out;
932
933 if (addr == NULL || addrlen != sizeof(struct sockaddr_dn))
934 goto out;
935 if (addr->sdn_family != AF_DECnet)
936 goto out;
937 if (addr->sdn_flags & SDF_WILD)
938 goto out;
939
940 if (sock_flag(sk, SOCK_ZAPPED)) {
941 err = dn_auto_bind(sk->sk_socket);
942 if (err)
943 goto out;
944 }
945
946 memcpy(&scp->peer, addr, sizeof(struct sockaddr_dn));
947
948 err = -EHOSTUNREACH;
949 memset(&fl, 0, sizeof(fl));
950 fl.oif = sk->sk_bound_dev_if;
951 fl.fld_dst = dn_saddr2dn(&scp->peer);
952 fl.fld_src = dn_saddr2dn(&scp->addr);
953 dn_sk_ports_copy(&fl, scp);
954 fl.proto = DNPROTO_NSP;
955 if (dn_route_output_sock(&sk->sk_dst_cache, &fl, sk, flags) < 0)
956 goto out;
957 sk->sk_route_caps = sk->sk_dst_cache->dev->features;
958 sock->state = SS_CONNECTING;
959 scp->state = DN_CI;
960 scp->segsize_loc = dst_metric(sk->sk_dst_cache, RTAX_ADVMSS);
961
962 dn_nsp_send_conninit(sk, NSP_CI);
963 err = -EINPROGRESS;
964 if (*timeo) {
965 err = dn_wait_run(sk, timeo);
966 }
967 out:
968 return err;
969 }
970
971 static int dn_connect(struct socket *sock, struct sockaddr *uaddr, int addrlen, int flags)
972 {
973 struct sockaddr_dn *addr = (struct sockaddr_dn *)uaddr;
974 struct sock *sk = sock->sk;
975 int err;
976 long timeo = sock_sndtimeo(sk, flags & O_NONBLOCK);
977
978 lock_sock(sk);
979 err = __dn_connect(sk, addr, addrlen, &timeo, 0);
980 release_sock(sk);
981
982 return err;
983 }
984
985 static inline int dn_check_state(struct sock *sk, struct sockaddr_dn *addr, int addrlen, long *timeo, int flags)
986 {
987 struct dn_scp *scp = DN_SK(sk);
988
989 switch(scp->state) {
990 case DN_RUN:
991 return 0;
992 case DN_CR:
993 return dn_confirm_accept(sk, timeo, sk->sk_allocation);
994 case DN_CI:
995 case DN_CC:
996 return dn_wait_run(sk, timeo);
997 case DN_O:
998 return __dn_connect(sk, addr, addrlen, timeo, flags);
999 }
1000
1001 return -EINVAL;
1002 }
1003
1004
1005 static void dn_access_copy(struct sk_buff *skb, struct accessdata_dn *acc)
1006 {
1007 unsigned char *ptr = skb->data;
1008
1009 acc->acc_userl = *ptr++;
1010 memcpy(&acc->acc_user, ptr, acc->acc_userl);
1011 ptr += acc->acc_userl;
1012
1013 acc->acc_passl = *ptr++;
1014 memcpy(&acc->acc_pass, ptr, acc->acc_passl);
1015 ptr += acc->acc_passl;
1016
1017 acc->acc_accl = *ptr++;
1018 memcpy(&acc->acc_acc, ptr, acc->acc_accl);
1019
1020 skb_pull(skb, acc->acc_accl + acc->acc_passl + acc->acc_userl + 3);
1021
1022 }
1023
1024 static void dn_user_copy(struct sk_buff *skb, struct optdata_dn *opt)
1025 {
1026 unsigned char *ptr = skb->data;
1027 u16 len = *ptr++; /* yes, it's 8bit on the wire */
1028
1029 BUG_ON(len > 16); /* we've checked the contents earlier */
1030 opt->opt_optl = cpu_to_le16(len);
1031 opt->opt_status = 0;
1032 memcpy(opt->opt_data, ptr, len);
1033 skb_pull(skb, len + 1);
1034 }
1035
1036 static struct sk_buff *dn_wait_for_connect(struct sock *sk, long *timeo)
1037 {
1038 DEFINE_WAIT(wait);
1039 struct sk_buff *skb = NULL;
1040 int err = 0;
1041
1042 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
1043 for(;;) {
1044 release_sock(sk);
1045 skb = skb_dequeue(&sk->sk_receive_queue);
1046 if (skb == NULL) {
1047 *timeo = schedule_timeout(*timeo);
1048 skb = skb_dequeue(&sk->sk_receive_queue);
1049 }
1050 lock_sock(sk);
1051 if (skb != NULL)
1052 break;
1053 err = -EINVAL;
1054 if (sk->sk_state != TCP_LISTEN)
1055 break;
1056 err = sock_intr_errno(*timeo);
1057 if (signal_pending(current))
1058 break;
1059 err = -EAGAIN;
1060 if (!*timeo)
1061 break;
1062 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
1063 }
1064 finish_wait(sk->sk_sleep, &wait);
1065
1066 return skb == NULL ? ERR_PTR(err) : skb;
1067 }
1068
1069 static int dn_accept(struct socket *sock, struct socket *newsock, int flags)
1070 {
1071 struct sock *sk = sock->sk, *newsk;
1072 struct sk_buff *skb = NULL;
1073 struct dn_skb_cb *cb;
1074 unsigned char menuver;
1075 int err = 0;
1076 unsigned char type;
1077 long timeo = sock_rcvtimeo(sk, flags & O_NONBLOCK);
1078
1079 lock_sock(sk);
1080
1081 if (sk->sk_state != TCP_LISTEN || DN_SK(sk)->state != DN_O) {
1082 release_sock(sk);
1083 return -EINVAL;
1084 }
1085
1086 skb = skb_dequeue(&sk->sk_receive_queue);
1087 if (skb == NULL) {
1088 skb = dn_wait_for_connect(sk, &timeo);
1089 if (IS_ERR(skb)) {
1090 release_sock(sk);
1091 return PTR_ERR(skb);
1092 }
1093 }
1094
1095 cb = DN_SKB_CB(skb);
1096 sk->sk_ack_backlog--;
1097 newsk = dn_alloc_sock(sock_net(sk), newsock, sk->sk_allocation);
1098 if (newsk == NULL) {
1099 release_sock(sk);
1100 kfree_skb(skb);
1101 return -ENOBUFS;
1102 }
1103 release_sock(sk);
1104
1105 dst_release(xchg(&newsk->sk_dst_cache, skb->dst));
1106 skb->dst = NULL;
1107
1108 DN_SK(newsk)->state = DN_CR;
1109 DN_SK(newsk)->addrrem = cb->src_port;
1110 DN_SK(newsk)->services_rem = cb->services;
1111 DN_SK(newsk)->info_rem = cb->info;
1112 DN_SK(newsk)->segsize_rem = cb->segsize;
1113 DN_SK(newsk)->accept_mode = DN_SK(sk)->accept_mode;
1114
1115 if (DN_SK(newsk)->segsize_rem < 230)
1116 DN_SK(newsk)->segsize_rem = 230;
1117
1118 if ((DN_SK(newsk)->services_rem & NSP_FC_MASK) == NSP_FC_NONE)
1119 DN_SK(newsk)->max_window = decnet_no_fc_max_cwnd;
1120
1121 newsk->sk_state = TCP_LISTEN;
1122 memcpy(&(DN_SK(newsk)->addr), &(DN_SK(sk)->addr), sizeof(struct sockaddr_dn));
1123
1124 /*
1125 * If we are listening on a wild socket, we don't want
1126 * the newly created socket on the wrong hash queue.
1127 */
1128 DN_SK(newsk)->addr.sdn_flags &= ~SDF_WILD;
1129
1130 skb_pull(skb, dn_username2sockaddr(skb->data, skb->len, &(DN_SK(newsk)->addr), &type));
1131 skb_pull(skb, dn_username2sockaddr(skb->data, skb->len, &(DN_SK(newsk)->peer), &type));
1132 *(__le16 *)(DN_SK(newsk)->peer.sdn_add.a_addr) = cb->src;
1133 *(__le16 *)(DN_SK(newsk)->addr.sdn_add.a_addr) = cb->dst;
1134
1135 menuver = *skb->data;
1136 skb_pull(skb, 1);
1137
1138 if (menuver & DN_MENUVER_ACC)
1139 dn_access_copy(skb, &(DN_SK(newsk)->accessdata));
1140
1141 if (menuver & DN_MENUVER_USR)
1142 dn_user_copy(skb, &(DN_SK(newsk)->conndata_in));
1143
1144 if (menuver & DN_MENUVER_PRX)
1145 DN_SK(newsk)->peer.sdn_flags |= SDF_PROXY;
1146
1147 if (menuver & DN_MENUVER_UIC)
1148 DN_SK(newsk)->peer.sdn_flags |= SDF_UICPROXY;
1149
1150 kfree_skb(skb);
1151
1152 memcpy(&(DN_SK(newsk)->conndata_out), &(DN_SK(sk)->conndata_out),
1153 sizeof(struct optdata_dn));
1154 memcpy(&(DN_SK(newsk)->discdata_out), &(DN_SK(sk)->discdata_out),
1155 sizeof(struct optdata_dn));
1156
1157 lock_sock(newsk);
1158 err = dn_hash_sock(newsk);
1159 if (err == 0) {
1160 sock_reset_flag(newsk, SOCK_ZAPPED);
1161 dn_send_conn_ack(newsk);
1162
1163 /*
1164 * Here we use sk->sk_allocation since although the conn conf is
1165 * for the newsk, the context is the old socket.
1166 */
1167 if (DN_SK(newsk)->accept_mode == ACC_IMMED)
1168 err = dn_confirm_accept(newsk, &timeo,
1169 sk->sk_allocation);
1170 }
1171 release_sock(newsk);
1172 return err;
1173 }
1174
1175
1176 static int dn_getname(struct socket *sock, struct sockaddr *uaddr,int *uaddr_len,int peer)
1177 {
1178 struct sockaddr_dn *sa = (struct sockaddr_dn *)uaddr;
1179 struct sock *sk = sock->sk;
1180 struct dn_scp *scp = DN_SK(sk);
1181
1182 *uaddr_len = sizeof(struct sockaddr_dn);
1183
1184 lock_sock(sk);
1185
1186 if (peer) {
1187 if ((sock->state != SS_CONNECTED &&
1188 sock->state != SS_CONNECTING) &&
1189 scp->accept_mode == ACC_IMMED) {
1190 release_sock(sk);
1191 return -ENOTCONN;
1192 }
1193
1194 memcpy(sa, &scp->peer, sizeof(struct sockaddr_dn));
1195 } else {
1196 memcpy(sa, &scp->addr, sizeof(struct sockaddr_dn));
1197 }
1198
1199 release_sock(sk);
1200
1201 return 0;
1202 }
1203
1204
1205 static unsigned int dn_poll(struct file *file, struct socket *sock, poll_table *wait)
1206 {
1207 struct sock *sk = sock->sk;
1208 struct dn_scp *scp = DN_SK(sk);
1209 int mask = datagram_poll(file, sock, wait);
1210
1211 if (!skb_queue_empty(&scp->other_receive_queue))
1212 mask |= POLLRDBAND;
1213
1214 return mask;
1215 }
1216
1217 static int dn_ioctl(struct socket *sock, unsigned int cmd, unsigned long arg)
1218 {
1219 struct sock *sk = sock->sk;
1220 struct dn_scp *scp = DN_SK(sk);
1221 int err = -EOPNOTSUPP;
1222 long amount = 0;
1223 struct sk_buff *skb;
1224 int val;
1225
1226 switch(cmd)
1227 {
1228 case SIOCGIFADDR:
1229 case SIOCSIFADDR:
1230 return dn_dev_ioctl(cmd, (void __user *)arg);
1231
1232 case SIOCATMARK:
1233 lock_sock(sk);
1234 val = !skb_queue_empty(&scp->other_receive_queue);
1235 if (scp->state != DN_RUN)
1236 val = -ENOTCONN;
1237 release_sock(sk);
1238 return val;
1239
1240 case TIOCOUTQ:
1241 amount = sk->sk_sndbuf - atomic_read(&sk->sk_wmem_alloc);
1242 if (amount < 0)
1243 amount = 0;
1244 err = put_user(amount, (int __user *)arg);
1245 break;
1246
1247 case TIOCINQ:
1248 lock_sock(sk);
1249 skb = skb_peek(&scp->other_receive_queue);
1250 if (skb) {
1251 amount = skb->len;
1252 } else {
1253 skb = sk->sk_receive_queue.next;
1254 for (;;) {
1255 if (skb ==
1256 (struct sk_buff *)&sk->sk_receive_queue)
1257 break;
1258 amount += skb->len;
1259 skb = skb->next;
1260 }
1261 }
1262 release_sock(sk);
1263 err = put_user(amount, (int __user *)arg);
1264 break;
1265
1266 default:
1267 err = -ENOIOCTLCMD;
1268 break;
1269 }
1270
1271 return err;
1272 }
1273
1274 static int dn_listen(struct socket *sock, int backlog)
1275 {
1276 struct sock *sk = sock->sk;
1277 int err = -EINVAL;
1278
1279 lock_sock(sk);
1280
1281 if (sock_flag(sk, SOCK_ZAPPED))
1282 goto out;
1283
1284 if ((DN_SK(sk)->state != DN_O) || (sk->sk_state == TCP_LISTEN))
1285 goto out;
1286
1287 sk->sk_max_ack_backlog = backlog;
1288 sk->sk_ack_backlog = 0;
1289 sk->sk_state = TCP_LISTEN;
1290 err = 0;
1291 dn_rehash_sock(sk);
1292
1293 out:
1294 release_sock(sk);
1295
1296 return err;
1297 }
1298
1299
1300 static int dn_shutdown(struct socket *sock, int how)
1301 {
1302 struct sock *sk = sock->sk;
1303 struct dn_scp *scp = DN_SK(sk);
1304 int err = -ENOTCONN;
1305
1306 lock_sock(sk);
1307
1308 if (sock->state == SS_UNCONNECTED)
1309 goto out;
1310
1311 err = 0;
1312 if (sock->state == SS_DISCONNECTING)
1313 goto out;
1314
1315 err = -EINVAL;
1316 if (scp->state == DN_O)
1317 goto out;
1318
1319 if (how != SHUTDOWN_MASK)
1320 goto out;
1321
1322 sk->sk_shutdown = how;
1323 dn_destroy_sock(sk);
1324 err = 0;
1325
1326 out:
1327 release_sock(sk);
1328
1329 return err;
1330 }
1331
1332 static int dn_setsockopt(struct socket *sock, int level, int optname, char __user *optval, int optlen)
1333 {
1334 struct sock *sk = sock->sk;
1335 int err;
1336
1337 lock_sock(sk);
1338 err = __dn_setsockopt(sock, level, optname, optval, optlen, 0);
1339 release_sock(sk);
1340
1341 return err;
1342 }
1343
1344 static int __dn_setsockopt(struct socket *sock, int level,int optname, char __user *optval, int optlen, int flags)
1345 {
1346 struct sock *sk = sock->sk;
1347 struct dn_scp *scp = DN_SK(sk);
1348 long timeo;
1349 union {
1350 struct optdata_dn opt;
1351 struct accessdata_dn acc;
1352 int mode;
1353 unsigned long win;
1354 int val;
1355 unsigned char services;
1356 unsigned char info;
1357 } u;
1358 int err;
1359
1360 if (optlen && !optval)
1361 return -EINVAL;
1362
1363 if (optlen > sizeof(u))
1364 return -EINVAL;
1365
1366 if (copy_from_user(&u, optval, optlen))
1367 return -EFAULT;
1368
1369 switch(optname) {
1370 case DSO_CONDATA:
1371 if (sock->state == SS_CONNECTED)
1372 return -EISCONN;
1373 if ((scp->state != DN_O) && (scp->state != DN_CR))
1374 return -EINVAL;
1375
1376 if (optlen != sizeof(struct optdata_dn))
1377 return -EINVAL;
1378
1379 if (le16_to_cpu(u.opt.opt_optl) > 16)
1380 return -EINVAL;
1381
1382 memcpy(&scp->conndata_out, &u.opt, optlen);
1383 break;
1384
1385 case DSO_DISDATA:
1386 if (sock->state != SS_CONNECTED && scp->accept_mode == ACC_IMMED)
1387 return -ENOTCONN;
1388
1389 if (optlen != sizeof(struct optdata_dn))
1390 return -EINVAL;
1391
1392 if (le16_to_cpu(u.opt.opt_optl) > 16)
1393 return -EINVAL;
1394
1395 memcpy(&scp->discdata_out, &u.opt, optlen);
1396 break;
1397
1398 case DSO_CONACCESS:
1399 if (sock->state == SS_CONNECTED)
1400 return -EISCONN;
1401 if (scp->state != DN_O)
1402 return -EINVAL;
1403
1404 if (optlen != sizeof(struct accessdata_dn))
1405 return -EINVAL;
1406
1407 if ((u.acc.acc_accl > DN_MAXACCL) ||
1408 (u.acc.acc_passl > DN_MAXACCL) ||
1409 (u.acc.acc_userl > DN_MAXACCL))
1410 return -EINVAL;
1411
1412 memcpy(&scp->accessdata, &u.acc, optlen);
1413 break;
1414
1415 case DSO_ACCEPTMODE:
1416 if (sock->state == SS_CONNECTED)
1417 return -EISCONN;
1418 if (scp->state != DN_O)
1419 return -EINVAL;
1420
1421 if (optlen != sizeof(int))
1422 return -EINVAL;
1423
1424 if ((u.mode != ACC_IMMED) && (u.mode != ACC_DEFER))
1425 return -EINVAL;
1426
1427 scp->accept_mode = (unsigned char)u.mode;
1428 break;
1429
1430 case DSO_CONACCEPT:
1431
1432 if (scp->state != DN_CR)
1433 return -EINVAL;
1434 timeo = sock_rcvtimeo(sk, 0);
1435 err = dn_confirm_accept(sk, &timeo, sk->sk_allocation);
1436 return err;
1437
1438 case DSO_CONREJECT:
1439
1440 if (scp->state != DN_CR)
1441 return -EINVAL;
1442
1443 scp->state = DN_DR;
1444 sk->sk_shutdown = SHUTDOWN_MASK;
1445 dn_nsp_send_disc(sk, 0x38, 0, sk->sk_allocation);
1446 break;
1447
1448 default:
1449 #ifdef CONFIG_NETFILTER
1450 return nf_setsockopt(sk, PF_DECnet, optname, optval, optlen);
1451 #endif
1452 case DSO_LINKINFO:
1453 case DSO_STREAM:
1454 case DSO_SEQPACKET:
1455 return -ENOPROTOOPT;
1456
1457 case DSO_MAXWINDOW:
1458 if (optlen != sizeof(unsigned long))
1459 return -EINVAL;
1460 if (u.win > NSP_MAX_WINDOW)
1461 u.win = NSP_MAX_WINDOW;
1462 if (u.win == 0)
1463 return -EINVAL;
1464 scp->max_window = u.win;
1465 if (scp->snd_window > u.win)
1466 scp->snd_window = u.win;
1467 break;
1468
1469 case DSO_NODELAY:
1470 if (optlen != sizeof(int))
1471 return -EINVAL;
1472 if (scp->nonagle == 2)
1473 return -EINVAL;
1474 scp->nonagle = (u.val == 0) ? 0 : 1;
1475 /* if (scp->nonagle == 1) { Push pending frames } */
1476 break;
1477
1478 case DSO_CORK:
1479 if (optlen != sizeof(int))
1480 return -EINVAL;
1481 if (scp->nonagle == 1)
1482 return -EINVAL;
1483 scp->nonagle = (u.val == 0) ? 0 : 2;
1484 /* if (scp->nonagle == 0) { Push pending frames } */
1485 break;
1486
1487 case DSO_SERVICES:
1488 if (optlen != sizeof(unsigned char))
1489 return -EINVAL;
1490 if ((u.services & ~NSP_FC_MASK) != 0x01)
1491 return -EINVAL;
1492 if ((u.services & NSP_FC_MASK) == NSP_FC_MASK)
1493 return -EINVAL;
1494 scp->services_loc = u.services;
1495 break;
1496
1497 case DSO_INFO:
1498 if (optlen != sizeof(unsigned char))
1499 return -EINVAL;
1500 if (u.info & 0xfc)
1501 return -EINVAL;
1502 scp->info_loc = u.info;
1503 break;
1504 }
1505
1506 return 0;
1507 }
1508
1509 static int dn_getsockopt(struct socket *sock, int level, int optname, char __user *optval, int __user *optlen)
1510 {
1511 struct sock *sk = sock->sk;
1512 int err;
1513
1514 lock_sock(sk);
1515 err = __dn_getsockopt(sock, level, optname, optval, optlen, 0);
1516 release_sock(sk);
1517
1518 return err;
1519 }
1520
1521 static int __dn_getsockopt(struct socket *sock, int level,int optname, char __user *optval,int __user *optlen, int flags)
1522 {
1523 struct sock *sk = sock->sk;
1524 struct dn_scp *scp = DN_SK(sk);
1525 struct linkinfo_dn link;
1526 unsigned int r_len;
1527 void *r_data = NULL;
1528 unsigned int val;
1529
1530 if(get_user(r_len , optlen))
1531 return -EFAULT;
1532
1533 switch(optname) {
1534 case DSO_CONDATA:
1535 if (r_len > sizeof(struct optdata_dn))
1536 r_len = sizeof(struct optdata_dn);
1537 r_data = &scp->conndata_in;
1538 break;
1539
1540 case DSO_DISDATA:
1541 if (r_len > sizeof(struct optdata_dn))
1542 r_len = sizeof(struct optdata_dn);
1543 r_data = &scp->discdata_in;
1544 break;
1545
1546 case DSO_CONACCESS:
1547 if (r_len > sizeof(struct accessdata_dn))
1548 r_len = sizeof(struct accessdata_dn);
1549 r_data = &scp->accessdata;
1550 break;
1551
1552 case DSO_ACCEPTMODE:
1553 if (r_len > sizeof(unsigned char))
1554 r_len = sizeof(unsigned char);
1555 r_data = &scp->accept_mode;
1556 break;
1557
1558 case DSO_LINKINFO:
1559 if (r_len > sizeof(struct linkinfo_dn))
1560 r_len = sizeof(struct linkinfo_dn);
1561
1562 switch(sock->state) {
1563 case SS_CONNECTING:
1564 link.idn_linkstate = LL_CONNECTING;
1565 break;
1566 case SS_DISCONNECTING:
1567 link.idn_linkstate = LL_DISCONNECTING;
1568 break;
1569 case SS_CONNECTED:
1570 link.idn_linkstate = LL_RUNNING;
1571 break;
1572 default:
1573 link.idn_linkstate = LL_INACTIVE;
1574 }
1575
1576 link.idn_segsize = scp->segsize_rem;
1577 r_data = &link;
1578 break;
1579
1580 default:
1581 #ifdef CONFIG_NETFILTER
1582 {
1583 int ret, len;
1584
1585 if(get_user(len, optlen))
1586 return -EFAULT;
1587
1588 ret = nf_getsockopt(sk, PF_DECnet, optname,
1589 optval, &len);
1590 if (ret >= 0)
1591 ret = put_user(len, optlen);
1592 return ret;
1593 }
1594 #endif
1595 case DSO_STREAM:
1596 case DSO_SEQPACKET:
1597 case DSO_CONACCEPT:
1598 case DSO_CONREJECT:
1599 return -ENOPROTOOPT;
1600
1601 case DSO_MAXWINDOW:
1602 if (r_len > sizeof(unsigned long))
1603 r_len = sizeof(unsigned long);
1604 r_data = &scp->max_window;
1605 break;
1606
1607 case DSO_NODELAY:
1608 if (r_len > sizeof(int))
1609 r_len = sizeof(int);
1610 val = (scp->nonagle == 1);
1611 r_data = &val;
1612 break;
1613
1614 case DSO_CORK:
1615 if (r_len > sizeof(int))
1616 r_len = sizeof(int);
1617 val = (scp->nonagle == 2);
1618 r_data = &val;
1619 break;
1620
1621 case DSO_SERVICES:
1622 if (r_len > sizeof(unsigned char))
1623 r_len = sizeof(unsigned char);
1624 r_data = &scp->services_rem;
1625 break;
1626
1627 case DSO_INFO:
1628 if (r_len > sizeof(unsigned char))
1629 r_len = sizeof(unsigned char);
1630 r_data = &scp->info_rem;
1631 break;
1632 }
1633
1634 if (r_data) {
1635 if (copy_to_user(optval, r_data, r_len))
1636 return -EFAULT;
1637 if (put_user(r_len, optlen))
1638 return -EFAULT;
1639 }
1640
1641 return 0;
1642 }
1643
1644
1645 static int dn_data_ready(struct sock *sk, struct sk_buff_head *q, int flags, int target)
1646 {
1647 struct sk_buff *skb = q->next;
1648 int len = 0;
1649
1650 if (flags & MSG_OOB)
1651 return !skb_queue_empty(q) ? 1 : 0;
1652
1653 while(skb != (struct sk_buff *)q) {
1654 struct dn_skb_cb *cb = DN_SKB_CB(skb);
1655 len += skb->len;
1656
1657 if (cb->nsp_flags & 0x40) {
1658 /* SOCK_SEQPACKET reads to EOM */
1659 if (sk->sk_type == SOCK_SEQPACKET)
1660 return 1;
1661 /* so does SOCK_STREAM unless WAITALL is specified */
1662 if (!(flags & MSG_WAITALL))
1663 return 1;
1664 }
1665
1666 /* minimum data length for read exceeded */
1667 if (len >= target)
1668 return 1;
1669
1670 skb = skb->next;
1671 }
1672
1673 return 0;
1674 }
1675
1676
1677 static int dn_recvmsg(struct kiocb *iocb, struct socket *sock,
1678 struct msghdr *msg, size_t size, int flags)
1679 {
1680 struct sock *sk = sock->sk;
1681 struct dn_scp *scp = DN_SK(sk);
1682 struct sk_buff_head *queue = &sk->sk_receive_queue;
1683 size_t target = size > 1 ? 1 : 0;
1684 size_t copied = 0;
1685 int rv = 0;
1686 struct sk_buff *skb, *nskb;
1687 struct dn_skb_cb *cb = NULL;
1688 unsigned char eor = 0;
1689 long timeo = sock_rcvtimeo(sk, flags & MSG_DONTWAIT);
1690
1691 lock_sock(sk);
1692
1693 if (sock_flag(sk, SOCK_ZAPPED)) {
1694 rv = -EADDRNOTAVAIL;
1695 goto out;
1696 }
1697
1698 if (sk->sk_shutdown & RCV_SHUTDOWN) {
1699 rv = 0;
1700 goto out;
1701 }
1702
1703 rv = dn_check_state(sk, NULL, 0, &timeo, flags);
1704 if (rv)
1705 goto out;
1706
1707 if (flags & ~(MSG_CMSG_COMPAT|MSG_PEEK|MSG_OOB|MSG_WAITALL|MSG_DONTWAIT|MSG_NOSIGNAL)) {
1708 rv = -EOPNOTSUPP;
1709 goto out;
1710 }
1711
1712 if (flags & MSG_OOB)
1713 queue = &scp->other_receive_queue;
1714
1715 if (flags & MSG_WAITALL)
1716 target = size;
1717
1718
1719 /*
1720 * See if there is data ready to read, sleep if there isn't
1721 */
1722 for(;;) {
1723 DEFINE_WAIT(wait);
1724
1725 if (sk->sk_err)
1726 goto out;
1727
1728 if (!skb_queue_empty(&scp->other_receive_queue)) {
1729 if (!(flags & MSG_OOB)) {
1730 msg->msg_flags |= MSG_OOB;
1731 if (!scp->other_report) {
1732 scp->other_report = 1;
1733 goto out;
1734 }
1735 }
1736 }
1737
1738 if (scp->state != DN_RUN)
1739 goto out;
1740
1741 if (signal_pending(current)) {
1742 rv = sock_intr_errno(timeo);
1743 goto out;
1744 }
1745
1746 if (dn_data_ready(sk, queue, flags, target))
1747 break;
1748
1749 if (flags & MSG_DONTWAIT) {
1750 rv = -EWOULDBLOCK;
1751 goto out;
1752 }
1753
1754 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
1755 set_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1756 sk_wait_event(sk, &timeo, dn_data_ready(sk, queue, flags, target));
1757 clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
1758 finish_wait(sk->sk_sleep, &wait);
1759 }
1760
1761 for(skb = queue->next; skb != (struct sk_buff *)queue; skb = nskb) {
1762 unsigned int chunk = skb->len;
1763 cb = DN_SKB_CB(skb);
1764
1765 if ((chunk + copied) > size)
1766 chunk = size - copied;
1767
1768 if (memcpy_toiovec(msg->msg_iov, skb->data, chunk)) {
1769 rv = -EFAULT;
1770 break;
1771 }
1772 copied += chunk;
1773
1774 if (!(flags & MSG_PEEK))
1775 skb_pull(skb, chunk);
1776
1777 eor = cb->nsp_flags & 0x40;
1778 nskb = skb->next;
1779
1780 if (skb->len == 0) {
1781 skb_unlink(skb, queue);
1782 kfree_skb(skb);
1783 /*
1784 * N.B. Don't refer to skb or cb after this point
1785 * in loop.
1786 */
1787 if ((scp->flowloc_sw == DN_DONTSEND) && !dn_congested(sk)) {
1788 scp->flowloc_sw = DN_SEND;
1789 dn_nsp_send_link(sk, DN_SEND, 0);
1790 }
1791 }
1792
1793 if (eor) {
1794 if (sk->sk_type == SOCK_SEQPACKET)
1795 break;
1796 if (!(flags & MSG_WAITALL))
1797 break;
1798 }
1799
1800 if (flags & MSG_OOB)
1801 break;
1802
1803 if (copied >= target)
1804 break;
1805 }
1806
1807 rv = copied;
1808
1809
1810 if (eor && (sk->sk_type == SOCK_SEQPACKET))
1811 msg->msg_flags |= MSG_EOR;
1812
1813 out:
1814 if (rv == 0)
1815 rv = (flags & MSG_PEEK) ? -sk->sk_err : sock_error(sk);
1816
1817 if ((rv >= 0) && msg->msg_name) {
1818 memcpy(msg->msg_name, &scp->peer, sizeof(struct sockaddr_dn));
1819 msg->msg_namelen = sizeof(struct sockaddr_dn);
1820 }
1821
1822 release_sock(sk);
1823
1824 return rv;
1825 }
1826
1827
1828 static inline int dn_queue_too_long(struct dn_scp *scp, struct sk_buff_head *queue, int flags)
1829 {
1830 unsigned char fctype = scp->services_rem & NSP_FC_MASK;
1831 if (skb_queue_len(queue) >= scp->snd_window)
1832 return 1;
1833 if (fctype != NSP_FC_NONE) {
1834 if (flags & MSG_OOB) {
1835 if (scp->flowrem_oth == 0)
1836 return 1;
1837 } else {
1838 if (scp->flowrem_dat == 0)
1839 return 1;
1840 }
1841 }
1842 return 0;
1843 }
1844
1845 /*
1846 * The DECnet spec requires that the "routing layer" accepts packets which
1847 * are at least 230 bytes in size. This excludes any headers which the NSP
1848 * layer might add, so we always assume that we'll be using the maximal
1849 * length header on data packets. The variation in length is due to the
1850 * inclusion (or not) of the two 16 bit acknowledgement fields so it doesn't
1851 * make much practical difference.
1852 */
1853 unsigned dn_mss_from_pmtu(struct net_device *dev, int mtu)
1854 {
1855 unsigned mss = 230 - DN_MAX_NSP_DATA_HEADER;
1856 if (dev) {
1857 struct dn_dev *dn_db = dev->dn_ptr;
1858 mtu -= LL_RESERVED_SPACE(dev);
1859 if (dn_db->use_long)
1860 mtu -= 21;
1861 else
1862 mtu -= 6;
1863 mtu -= DN_MAX_NSP_DATA_HEADER;
1864 } else {
1865 /*
1866 * 21 = long header, 16 = guess at MAC header length
1867 */
1868 mtu -= (21 + DN_MAX_NSP_DATA_HEADER + 16);
1869 }
1870 if (mtu > mss)
1871 mss = mtu;
1872 return mss;
1873 }
1874
1875 static inline unsigned int dn_current_mss(struct sock *sk, int flags)
1876 {
1877 struct dst_entry *dst = __sk_dst_get(sk);
1878 struct dn_scp *scp = DN_SK(sk);
1879 int mss_now = min_t(int, scp->segsize_loc, scp->segsize_rem);
1880
1881 /* Other data messages are limited to 16 bytes per packet */
1882 if (flags & MSG_OOB)
1883 return 16;
1884
1885 /* This works out the maximum size of segment we can send out */
1886 if (dst) {
1887 u32 mtu = dst_mtu(dst);
1888 mss_now = min_t(int, dn_mss_from_pmtu(dst->dev, mtu), mss_now);
1889 }
1890
1891 return mss_now;
1892 }
1893
1894 /*
1895 * N.B. We get the timeout wrong here, but then we always did get it
1896 * wrong before and this is another step along the road to correcting
1897 * it. It ought to get updated each time we pass through the routine,
1898 * but in practise it probably doesn't matter too much for now.
1899 */
1900 static inline struct sk_buff *dn_alloc_send_pskb(struct sock *sk,
1901 unsigned long datalen, int noblock,
1902 int *errcode)
1903 {
1904 struct sk_buff *skb = sock_alloc_send_skb(sk, datalen,
1905 noblock, errcode);
1906 if (skb) {
1907 skb->protocol = htons(ETH_P_DNA_RT);
1908 skb->pkt_type = PACKET_OUTGOING;
1909 }
1910 return skb;
1911 }
1912
1913 static int dn_sendmsg(struct kiocb *iocb, struct socket *sock,
1914 struct msghdr *msg, size_t size)
1915 {
1916 struct sock *sk = sock->sk;
1917 struct dn_scp *scp = DN_SK(sk);
1918 size_t mss;
1919 struct sk_buff_head *queue = &scp->data_xmit_queue;
1920 int flags = msg->msg_flags;
1921 int err = 0;
1922 size_t sent = 0;
1923 int addr_len = msg->msg_namelen;
1924 struct sockaddr_dn *addr = (struct sockaddr_dn *)msg->msg_name;
1925 struct sk_buff *skb = NULL;
1926 struct dn_skb_cb *cb;
1927 size_t len;
1928 unsigned char fctype;
1929 long timeo;
1930
1931 if (flags & ~(MSG_TRYHARD|MSG_OOB|MSG_DONTWAIT|MSG_EOR|MSG_NOSIGNAL|MSG_MORE|MSG_CMSG_COMPAT))
1932 return -EOPNOTSUPP;
1933
1934 if (addr_len && (addr_len != sizeof(struct sockaddr_dn)))
1935 return -EINVAL;
1936
1937 lock_sock(sk);
1938 timeo = sock_sndtimeo(sk, flags & MSG_DONTWAIT);
1939 /*
1940 * The only difference between stream sockets and sequenced packet
1941 * sockets is that the stream sockets always behave as if MSG_EOR
1942 * has been set.
1943 */
1944 if (sock->type == SOCK_STREAM) {
1945 if (flags & MSG_EOR) {
1946 err = -EINVAL;
1947 goto out;
1948 }
1949 flags |= MSG_EOR;
1950 }
1951
1952
1953 err = dn_check_state(sk, addr, addr_len, &timeo, flags);
1954 if (err)
1955 goto out_err;
1956
1957 if (sk->sk_shutdown & SEND_SHUTDOWN) {
1958 err = -EPIPE;
1959 if (!(flags & MSG_NOSIGNAL))
1960 send_sig(SIGPIPE, current, 0);
1961 goto out_err;
1962 }
1963
1964 if ((flags & MSG_TRYHARD) && sk->sk_dst_cache)
1965 dst_negative_advice(&sk->sk_dst_cache);
1966
1967 mss = scp->segsize_rem;
1968 fctype = scp->services_rem & NSP_FC_MASK;
1969
1970 mss = dn_current_mss(sk, flags);
1971
1972 if (flags & MSG_OOB) {
1973 queue = &scp->other_xmit_queue;
1974 if (size > mss) {
1975 err = -EMSGSIZE;
1976 goto out;
1977 }
1978 }
1979
1980 scp->persist_fxn = dn_nsp_xmit_timeout;
1981
1982 while(sent < size) {
1983 err = sock_error(sk);
1984 if (err)
1985 goto out;
1986
1987 if (signal_pending(current)) {
1988 err = sock_intr_errno(timeo);
1989 goto out;
1990 }
1991
1992 /*
1993 * Calculate size that we wish to send.
1994 */
1995 len = size - sent;
1996
1997 if (len > mss)
1998 len = mss;
1999
2000 /*
2001 * Wait for queue size to go down below the window
2002 * size.
2003 */
2004 if (dn_queue_too_long(scp, queue, flags)) {
2005 DEFINE_WAIT(wait);
2006
2007 if (flags & MSG_DONTWAIT) {
2008 err = -EWOULDBLOCK;
2009 goto out;
2010 }
2011
2012 prepare_to_wait(sk->sk_sleep, &wait, TASK_INTERRUPTIBLE);
2013 set_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
2014 sk_wait_event(sk, &timeo,
2015 !dn_queue_too_long(scp, queue, flags));
2016 clear_bit(SOCK_ASYNC_WAITDATA, &sk->sk_socket->flags);
2017 finish_wait(sk->sk_sleep, &wait);
2018 continue;
2019 }
2020
2021 /*
2022 * Get a suitably sized skb.
2023 * 64 is a bit of a hack really, but its larger than any
2024 * link-layer headers and has served us well as a good
2025 * guess as to their real length.
2026 */
2027 skb = dn_alloc_send_pskb(sk, len + 64 + DN_MAX_NSP_DATA_HEADER,
2028 flags & MSG_DONTWAIT, &err);
2029
2030 if (err)
2031 break;
2032
2033 if (!skb)
2034 continue;
2035
2036 cb = DN_SKB_CB(skb);
2037
2038 skb_reserve(skb, 64 + DN_MAX_NSP_DATA_HEADER);
2039
2040 if (memcpy_fromiovec(skb_put(skb, len), msg->msg_iov, len)) {
2041 err = -EFAULT;
2042 goto out;
2043 }
2044
2045 if (flags & MSG_OOB) {
2046 cb->nsp_flags = 0x30;
2047 if (fctype != NSP_FC_NONE)
2048 scp->flowrem_oth--;
2049 } else {
2050 cb->nsp_flags = 0x00;
2051 if (scp->seg_total == 0)
2052 cb->nsp_flags |= 0x20;
2053
2054 scp->seg_total += len;
2055
2056 if (((sent + len) == size) && (flags & MSG_EOR)) {
2057 cb->nsp_flags |= 0x40;
2058 scp->seg_total = 0;
2059 if (fctype == NSP_FC_SCMC)
2060 scp->flowrem_dat--;
2061 }
2062 if (fctype == NSP_FC_SRC)
2063 scp->flowrem_dat--;
2064 }
2065
2066 sent += len;
2067 dn_nsp_queue_xmit(sk, skb, sk->sk_allocation, flags & MSG_OOB);
2068 skb = NULL;
2069
2070 scp->persist = dn_nsp_persist(sk);
2071
2072 }
2073 out:
2074
2075 kfree_skb(skb);
2076
2077 release_sock(sk);
2078
2079 return sent ? sent : err;
2080
2081 out_err:
2082 err = sk_stream_error(sk, flags, err);
2083 release_sock(sk);
2084 return err;
2085 }
2086
2087 static int dn_device_event(struct notifier_block *this, unsigned long event,
2088 void *ptr)
2089 {
2090 struct net_device *dev = (struct net_device *)ptr;
2091
2092 if (!net_eq(dev_net(dev), &init_net))
2093 return NOTIFY_DONE;
2094
2095 switch(event) {
2096 case NETDEV_UP:
2097 dn_dev_up(dev);
2098 break;
2099 case NETDEV_DOWN:
2100 dn_dev_down(dev);
2101 break;
2102 default:
2103 break;
2104 }
2105
2106 return NOTIFY_DONE;
2107 }
2108
2109 static struct notifier_block dn_dev_notifier = {
2110 .notifier_call = dn_device_event,
2111 };
2112
2113 extern int dn_route_rcv(struct sk_buff *, struct net_device *, struct packet_type *, struct net_device *);
2114
2115 static struct packet_type dn_dix_packet_type __read_mostly = {
2116 .type = cpu_to_be16(ETH_P_DNA_RT),
2117 .func = dn_route_rcv,
2118 };
2119
2120 #ifdef CONFIG_PROC_FS
2121 struct dn_iter_state {
2122 int bucket;
2123 };
2124
2125 static struct sock *dn_socket_get_first(struct seq_file *seq)
2126 {
2127 struct dn_iter_state *state = seq->private;
2128 struct sock *n = NULL;
2129
2130 for(state->bucket = 0;
2131 state->bucket < DN_SK_HASH_SIZE;
2132 ++state->bucket) {
2133 n = sk_head(&dn_sk_hash[state->bucket]);
2134 if (n)
2135 break;
2136 }
2137
2138 return n;
2139 }
2140
2141 static struct sock *dn_socket_get_next(struct seq_file *seq,
2142 struct sock *n)
2143 {
2144 struct dn_iter_state *state = seq->private;
2145
2146 n = sk_next(n);
2147 try_again:
2148 if (n)
2149 goto out;
2150 if (++state->bucket >= DN_SK_HASH_SIZE)
2151 goto out;
2152 n = sk_head(&dn_sk_hash[state->bucket]);
2153 goto try_again;
2154 out:
2155 return n;
2156 }
2157
2158 static struct sock *socket_get_idx(struct seq_file *seq, loff_t *pos)
2159 {
2160 struct sock *sk = dn_socket_get_first(seq);
2161
2162 if (sk) {
2163 while(*pos && (sk = dn_socket_get_next(seq, sk)))
2164 --*pos;
2165 }
2166 return *pos ? NULL : sk;
2167 }
2168
2169 static void *dn_socket_get_idx(struct seq_file *seq, loff_t pos)
2170 {
2171 void *rc;
2172 read_lock_bh(&dn_hash_lock);
2173 rc = socket_get_idx(seq, &pos);
2174 if (!rc) {
2175 read_unlock_bh(&dn_hash_lock);
2176 }
2177 return rc;
2178 }
2179
2180 static void *dn_socket_seq_start(struct seq_file *seq, loff_t *pos)
2181 {
2182 return *pos ? dn_socket_get_idx(seq, *pos - 1) : SEQ_START_TOKEN;
2183 }
2184
2185 static void *dn_socket_seq_next(struct seq_file *seq, void *v, loff_t *pos)
2186 {
2187 void *rc;
2188
2189 if (v == SEQ_START_TOKEN) {
2190 rc = dn_socket_get_idx(seq, 0);
2191 goto out;
2192 }
2193
2194 rc = dn_socket_get_next(seq, v);
2195 if (rc)
2196 goto out;
2197 read_unlock_bh(&dn_hash_lock);
2198 out:
2199 ++*pos;
2200 return rc;
2201 }
2202
2203 static void dn_socket_seq_stop(struct seq_file *seq, void *v)
2204 {
2205 if (v && v != SEQ_START_TOKEN)
2206 read_unlock_bh(&dn_hash_lock);
2207 }
2208
2209 #define IS_NOT_PRINTABLE(x) ((x) < 32 || (x) > 126)
2210
2211 static void dn_printable_object(struct sockaddr_dn *dn, unsigned char *buf)
2212 {
2213 int i;
2214
2215 switch (le16_to_cpu(dn->sdn_objnamel)) {
2216 case 0:
2217 sprintf(buf, "%d", dn->sdn_objnum);
2218 break;
2219 default:
2220 for (i = 0; i < le16_to_cpu(dn->sdn_objnamel); i++) {
2221 buf[i] = dn->sdn_objname[i];
2222 if (IS_NOT_PRINTABLE(buf[i]))
2223 buf[i] = '.';
2224 }
2225 buf[i] = 0;
2226 }
2227 }
2228
2229 static char *dn_state2asc(unsigned char state)
2230 {
2231 switch(state) {
2232 case DN_O:
2233 return "OPEN";
2234 case DN_CR:
2235 return " CR";
2236 case DN_DR:
2237 return " DR";
2238 case DN_DRC:
2239 return " DRC";
2240 case DN_CC:
2241 return " CC";
2242 case DN_CI:
2243 return " CI";
2244 case DN_NR:
2245 return " NR";
2246 case DN_NC:
2247 return " NC";
2248 case DN_CD:
2249 return " CD";
2250 case DN_RJ:
2251 return " RJ";
2252 case DN_RUN:
2253 return " RUN";
2254 case DN_DI:
2255 return " DI";
2256 case DN_DIC:
2257 return " DIC";
2258 case DN_DN:
2259 return " DN";
2260 case DN_CL:
2261 return " CL";
2262 case DN_CN:
2263 return " CN";
2264 }
2265
2266 return "????";
2267 }
2268
2269 static inline void dn_socket_format_entry(struct seq_file *seq, struct sock *sk)
2270 {
2271 struct dn_scp *scp = DN_SK(sk);
2272 char buf1[DN_ASCBUF_LEN];
2273 char buf2[DN_ASCBUF_LEN];
2274 char local_object[DN_MAXOBJL+3];
2275 char remote_object[DN_MAXOBJL+3];
2276
2277 dn_printable_object(&scp->addr, local_object);
2278 dn_printable_object(&scp->peer, remote_object);
2279
2280 seq_printf(seq,
2281 "%6s/%04X %04d:%04d %04d:%04d %01d %-16s "
2282 "%6s/%04X %04d:%04d %04d:%04d %01d %-16s %4s %s\n",
2283 dn_addr2asc(le16_to_cpu(dn_saddr2dn(&scp->addr)), buf1),
2284 scp->addrloc,
2285 scp->numdat,
2286 scp->numoth,
2287 scp->ackxmt_dat,
2288 scp->ackxmt_oth,
2289 scp->flowloc_sw,
2290 local_object,
2291 dn_addr2asc(le16_to_cpu(dn_saddr2dn(&scp->peer)), buf2),
2292 scp->addrrem,
2293 scp->numdat_rcv,
2294 scp->numoth_rcv,
2295 scp->ackrcv_dat,
2296 scp->ackrcv_oth,
2297 scp->flowrem_sw,
2298 remote_object,
2299 dn_state2asc(scp->state),
2300 ((scp->accept_mode == ACC_IMMED) ? "IMMED" : "DEFER"));
2301 }
2302
2303 static int dn_socket_seq_show(struct seq_file *seq, void *v)
2304 {
2305 if (v == SEQ_START_TOKEN) {
2306 seq_puts(seq, "Local Remote\n");
2307 } else {
2308 dn_socket_format_entry(seq, v);
2309 }
2310 return 0;
2311 }
2312
2313 static const struct seq_operations dn_socket_seq_ops = {
2314 .start = dn_socket_seq_start,
2315 .next = dn_socket_seq_next,
2316 .stop = dn_socket_seq_stop,
2317 .show = dn_socket_seq_show,
2318 };
2319
2320 static int dn_socket_seq_open(struct inode *inode, struct file *file)
2321 {
2322 return seq_open_private(file, &dn_socket_seq_ops,
2323 sizeof(struct dn_iter_state));
2324 }
2325
2326 static const struct file_operations dn_socket_seq_fops = {
2327 .owner = THIS_MODULE,
2328 .open = dn_socket_seq_open,
2329 .read = seq_read,
2330 .llseek = seq_lseek,
2331 .release = seq_release_private,
2332 };
2333 #endif
2334
2335 static struct net_proto_family dn_family_ops = {
2336 .family = AF_DECnet,
2337 .create = dn_create,
2338 .owner = THIS_MODULE,
2339 };
2340
2341 static const struct proto_ops dn_proto_ops = {
2342 .family = AF_DECnet,
2343 .owner = THIS_MODULE,
2344 .release = dn_release,
2345 .bind = dn_bind,
2346 .connect = dn_connect,
2347 .socketpair = sock_no_socketpair,
2348 .accept = dn_accept,
2349 .getname = dn_getname,
2350 .poll = dn_poll,
2351 .ioctl = dn_ioctl,
2352 .listen = dn_listen,
2353 .shutdown = dn_shutdown,
2354 .setsockopt = dn_setsockopt,
2355 .getsockopt = dn_getsockopt,
2356 .sendmsg = dn_sendmsg,
2357 .recvmsg = dn_recvmsg,
2358 .mmap = sock_no_mmap,
2359 .sendpage = sock_no_sendpage,
2360 };
2361
2362 void dn_register_sysctl(void);
2363 void dn_unregister_sysctl(void);
2364
2365 MODULE_DESCRIPTION("The Linux DECnet Network Protocol");
2366 MODULE_AUTHOR("Linux DECnet Project Team");
2367 MODULE_LICENSE("GPL");
2368 MODULE_ALIAS_NETPROTO(PF_DECnet);
2369
2370 static char banner[] __initdata = KERN_INFO "NET4: DECnet for Linux: V.2.5.68s (C) 1995-2003 Linux DECnet Project Team\n";
2371
2372 static int __init decnet_init(void)
2373 {
2374 int rc;
2375
2376 printk(banner);
2377
2378 rc = proto_register(&dn_proto, 1);
2379 if (rc != 0)
2380 goto out;
2381
2382 dn_neigh_init();
2383 dn_dev_init();
2384 dn_route_init();
2385 dn_fib_init();
2386
2387 sock_register(&dn_family_ops);
2388 dev_add_pack(&dn_dix_packet_type);
2389 register_netdevice_notifier(&dn_dev_notifier);
2390
2391 proc_net_fops_create(&init_net, "decnet", S_IRUGO, &dn_socket_seq_fops);
2392 dn_register_sysctl();
2393 out:
2394 return rc;
2395
2396 }
2397 module_init(decnet_init);
2398
2399 /*
2400 * Prevent DECnet module unloading until its fixed properly.
2401 * Requires an audit of the code to check for memory leaks and
2402 * initialisation problems etc.
2403 */
2404 #if 0
2405 static void __exit decnet_exit(void)
2406 {
2407 sock_unregister(AF_DECnet);
2408 rtnl_unregister_all(PF_DECnet);
2409 dev_remove_pack(&dn_dix_packet_type);
2410
2411 dn_unregister_sysctl();
2412
2413 unregister_netdevice_notifier(&dn_dev_notifier);
2414
2415 dn_route_cleanup();
2416 dn_dev_cleanup();
2417 dn_neigh_cleanup();
2418 dn_fib_cleanup();
2419
2420 proc_net_remove(&init_net, "decnet");
2421
2422 proto_unregister(&dn_proto);
2423 }
2424 module_exit(decnet_exit);
2425 #endif