[NETFILTER]: arp_tables: move counter allocation to seperate function
[GitHub/moto-9609/android_kernel_motorola_exynos9610.git] / net / ipv4 / netfilter / arp_tables.c
CommitLineData
1da177e4
LT
1/*
2 * Packet matching code for ARP packets.
3 *
4 * Based heavily, if not almost entirely, upon ip_tables.c framework.
5 *
6 * Some ARP specific bits are:
7 *
8 * Copyright (C) 2002 David S. Miller (davem@redhat.com)
9 *
10 */
11
1da177e4
LT
12#include <linux/kernel.h>
13#include <linux/skbuff.h>
14#include <linux/netdevice.h>
4fc268d2 15#include <linux/capability.h>
1da177e4
LT
16#include <linux/if_arp.h>
17#include <linux/kmod.h>
18#include <linux/vmalloc.h>
19#include <linux/proc_fs.h>
20#include <linux/module.h>
21#include <linux/init.h>
22
23#include <asm/uaccess.h>
57b47a53 24#include <linux/mutex.h>
1da177e4 25
2e4e6a17 26#include <linux/netfilter/x_tables.h>
1da177e4
LT
27#include <linux/netfilter_arp/arp_tables.h>
28
29MODULE_LICENSE("GPL");
30MODULE_AUTHOR("David S. Miller <davem@redhat.com>");
31MODULE_DESCRIPTION("arptables core");
32
33/*#define DEBUG_ARP_TABLES*/
34/*#define DEBUG_ARP_TABLES_USER*/
35
36#ifdef DEBUG_ARP_TABLES
37#define dprintf(format, args...) printk(format , ## args)
38#else
39#define dprintf(format, args...)
40#endif
41
42#ifdef DEBUG_ARP_TABLES_USER
43#define duprintf(format, args...) printk(format , ## args)
44#else
45#define duprintf(format, args...)
46#endif
47
48#ifdef CONFIG_NETFILTER_DEBUG
49#define ARP_NF_ASSERT(x) \
50do { \
51 if (!(x)) \
52 printk("ARP_NF_ASSERT: %s:%s:%u\n", \
53 __FUNCTION__, __FILE__, __LINE__); \
54} while(0)
55#else
56#define ARP_NF_ASSERT(x)
57#endif
1da177e4 58
1da177e4
LT
59static inline int arp_devaddr_compare(const struct arpt_devaddr_info *ap,
60 char *hdr_addr, int len)
61{
62 int i, ret;
63
64 if (len > ARPT_DEV_ADDR_LEN_MAX)
65 len = ARPT_DEV_ADDR_LEN_MAX;
66
67 ret = 0;
68 for (i = 0; i < len; i++)
69 ret |= (hdr_addr[i] ^ ap->addr[i]) & ap->mask[i];
70
71 return (ret != 0);
72}
73
74/* Returns whether packet matches rule or not. */
75static inline int arp_packet_match(const struct arphdr *arphdr,
76 struct net_device *dev,
77 const char *indev,
78 const char *outdev,
79 const struct arpt_arp *arpinfo)
80{
81 char *arpptr = (char *)(arphdr + 1);
82 char *src_devaddr, *tgt_devaddr;
59b8bfd8 83 __be32 src_ipaddr, tgt_ipaddr;
1da177e4
LT
84 int i, ret;
85
86#define FWINV(bool,invflg) ((bool) ^ !!(arpinfo->invflags & invflg))
87
88 if (FWINV((arphdr->ar_op & arpinfo->arpop_mask) != arpinfo->arpop,
89 ARPT_INV_ARPOP)) {
90 dprintf("ARP operation field mismatch.\n");
91 dprintf("ar_op: %04x info->arpop: %04x info->arpop_mask: %04x\n",
92 arphdr->ar_op, arpinfo->arpop, arpinfo->arpop_mask);
93 return 0;
94 }
95
96 if (FWINV((arphdr->ar_hrd & arpinfo->arhrd_mask) != arpinfo->arhrd,
97 ARPT_INV_ARPHRD)) {
98 dprintf("ARP hardware address format mismatch.\n");
99 dprintf("ar_hrd: %04x info->arhrd: %04x info->arhrd_mask: %04x\n",
100 arphdr->ar_hrd, arpinfo->arhrd, arpinfo->arhrd_mask);
101 return 0;
102 }
103
104 if (FWINV((arphdr->ar_pro & arpinfo->arpro_mask) != arpinfo->arpro,
105 ARPT_INV_ARPPRO)) {
106 dprintf("ARP protocol address format mismatch.\n");
107 dprintf("ar_pro: %04x info->arpro: %04x info->arpro_mask: %04x\n",
108 arphdr->ar_pro, arpinfo->arpro, arpinfo->arpro_mask);
109 return 0;
110 }
111
112 if (FWINV((arphdr->ar_hln & arpinfo->arhln_mask) != arpinfo->arhln,
113 ARPT_INV_ARPHLN)) {
114 dprintf("ARP hardware address length mismatch.\n");
115 dprintf("ar_hln: %02x info->arhln: %02x info->arhln_mask: %02x\n",
116 arphdr->ar_hln, arpinfo->arhln, arpinfo->arhln_mask);
117 return 0;
118 }
119
120 src_devaddr = arpptr;
121 arpptr += dev->addr_len;
122 memcpy(&src_ipaddr, arpptr, sizeof(u32));
123 arpptr += sizeof(u32);
124 tgt_devaddr = arpptr;
125 arpptr += dev->addr_len;
126 memcpy(&tgt_ipaddr, arpptr, sizeof(u32));
127
128 if (FWINV(arp_devaddr_compare(&arpinfo->src_devaddr, src_devaddr, dev->addr_len),
129 ARPT_INV_SRCDEVADDR) ||
130 FWINV(arp_devaddr_compare(&arpinfo->tgt_devaddr, tgt_devaddr, dev->addr_len),
131 ARPT_INV_TGTDEVADDR)) {
132 dprintf("Source or target device address mismatch.\n");
133
134 return 0;
135 }
136
137 if (FWINV((src_ipaddr & arpinfo->smsk.s_addr) != arpinfo->src.s_addr,
138 ARPT_INV_SRCIP) ||
139 FWINV(((tgt_ipaddr & arpinfo->tmsk.s_addr) != arpinfo->tgt.s_addr),
140 ARPT_INV_TGTIP)) {
141 dprintf("Source or target IP address mismatch.\n");
142
143 dprintf("SRC: %u.%u.%u.%u. Mask: %u.%u.%u.%u. Target: %u.%u.%u.%u.%s\n",
144 NIPQUAD(src_ipaddr),
145 NIPQUAD(arpinfo->smsk.s_addr),
146 NIPQUAD(arpinfo->src.s_addr),
147 arpinfo->invflags & ARPT_INV_SRCIP ? " (INV)" : "");
148 dprintf("TGT: %u.%u.%u.%u Mask: %u.%u.%u.%u Target: %u.%u.%u.%u.%s\n",
149 NIPQUAD(tgt_ipaddr),
150 NIPQUAD(arpinfo->tmsk.s_addr),
151 NIPQUAD(arpinfo->tgt.s_addr),
152 arpinfo->invflags & ARPT_INV_TGTIP ? " (INV)" : "");
153 return 0;
154 }
155
156 /* Look for ifname matches. */
157 for (i = 0, ret = 0; i < IFNAMSIZ; i++) {
158 ret |= (indev[i] ^ arpinfo->iniface[i])
159 & arpinfo->iniface_mask[i];
160 }
161
162 if (FWINV(ret != 0, ARPT_INV_VIA_IN)) {
163 dprintf("VIA in mismatch (%s vs %s).%s\n",
164 indev, arpinfo->iniface,
165 arpinfo->invflags&ARPT_INV_VIA_IN ?" (INV)":"");
166 return 0;
167 }
168
49688c84
DM
169 for (i = 0, ret = 0; i < IFNAMSIZ; i++) {
170 ret |= (outdev[i] ^ arpinfo->outiface[i])
171 & arpinfo->outiface_mask[i];
1da177e4
LT
172 }
173
174 if (FWINV(ret != 0, ARPT_INV_VIA_OUT)) {
175 dprintf("VIA out mismatch (%s vs %s).%s\n",
176 outdev, arpinfo->outiface,
177 arpinfo->invflags&ARPT_INV_VIA_OUT ?" (INV)":"");
178 return 0;
179 }
180
181 return 1;
182}
183
184static inline int arp_checkentry(const struct arpt_arp *arp)
185{
186 if (arp->flags & ~ARPT_F_MASK) {
187 duprintf("Unknown flag bits set: %08X\n",
188 arp->flags & ~ARPT_F_MASK);
189 return 0;
190 }
191 if (arp->invflags & ~ARPT_INV_MASK) {
192 duprintf("Unknown invflag bits set: %08X\n",
193 arp->invflags & ~ARPT_INV_MASK);
194 return 0;
195 }
196
197 return 1;
198}
199
3db05fea 200static unsigned int arpt_error(struct sk_buff *skb,
1da177e4
LT
201 const struct net_device *in,
202 const struct net_device *out,
2e4e6a17 203 unsigned int hooknum,
c4986734 204 const struct xt_target *target,
fe1cb108 205 const void *targinfo)
1da177e4
LT
206{
207 if (net_ratelimit())
208 printk("arp_tables: error: '%s'\n", (char *)targinfo);
209
210 return NF_DROP;
211}
212
213static inline struct arpt_entry *get_entry(void *base, unsigned int offset)
214{
215 return (struct arpt_entry *)(base + offset);
216}
217
3db05fea 218unsigned int arpt_do_table(struct sk_buff *skb,
1da177e4
LT
219 unsigned int hook,
220 const struct net_device *in,
221 const struct net_device *out,
fe1cb108 222 struct arpt_table *table)
1da177e4
LT
223{
224 static const char nulldevname[IFNAMSIZ];
225 unsigned int verdict = NF_DROP;
226 struct arphdr *arp;
cff533ac 227 bool hotdrop = false;
1da177e4
LT
228 struct arpt_entry *e, *back;
229 const char *indev, *outdev;
230 void *table_base;
e0b7cde9 231 struct xt_table_info *private;
1da177e4
LT
232
233 /* ARP header, plus 2 device addresses, plus 2 IP addresses. */
3db05fea
HX
234 if (!pskb_may_pull(skb, (sizeof(struct arphdr) +
235 (2 * skb->dev->addr_len) +
236 (2 * sizeof(u32)))))
1da177e4
LT
237 return NF_DROP;
238
239 indev = in ? in->name : nulldevname;
240 outdev = out ? out->name : nulldevname;
241
242 read_lock_bh(&table->lock);
e0b7cde9 243 private = table->private;
2e4e6a17
HW
244 table_base = (void *)private->entries[smp_processor_id()];
245 e = get_entry(table_base, private->hook_entry[hook]);
246 back = get_entry(table_base, private->underflow[hook]);
1da177e4 247
3db05fea 248 arp = arp_hdr(skb);
1da177e4 249 do {
3db05fea 250 if (arp_packet_match(arp, skb->dev, indev, outdev, &e->arp)) {
1da177e4
LT
251 struct arpt_entry_target *t;
252 int hdr_len;
253
254 hdr_len = sizeof(*arp) + (2 * sizeof(struct in_addr)) +
3db05fea 255 (2 * skb->dev->addr_len);
1da177e4
LT
256 ADD_COUNTER(e->counters, hdr_len, 1);
257
258 t = arpt_get_target(e);
259
260 /* Standard target? */
261 if (!t->u.kernel.target->target) {
262 int v;
263
264 v = ((struct arpt_standard_target *)t)->verdict;
265 if (v < 0) {
266 /* Pop from stack? */
267 if (v != ARPT_RETURN) {
268 verdict = (unsigned)(-v) - 1;
269 break;
270 }
271 e = back;
272 back = get_entry(table_base,
273 back->comefrom);
274 continue;
275 }
276 if (table_base + v
277 != (void *)e + e->next_offset) {
278 /* Save old back ptr in next entry */
279 struct arpt_entry *next
280 = (void *)e + e->next_offset;
281 next->comefrom =
282 (void *)back - table_base;
283
284 /* set back pointer to next entry */
285 back = next;
286 }
287
288 e = get_entry(table_base, v);
289 } else {
290 /* Targets which reenter must return
291 * abs. verdicts
292 */
3db05fea 293 verdict = t->u.kernel.target->target(skb,
1da177e4 294 in, out,
2e4e6a17 295 hook,
1c524830 296 t->u.kernel.target,
fe1cb108 297 t->data);
1da177e4
LT
298
299 /* Target might have changed stuff. */
3db05fea 300 arp = arp_hdr(skb);
1da177e4
LT
301
302 if (verdict == ARPT_CONTINUE)
303 e = (void *)e + e->next_offset;
304 else
305 /* Verdict */
306 break;
307 }
308 } else {
309 e = (void *)e + e->next_offset;
310 }
311 } while (!hotdrop);
312 read_unlock_bh(&table->lock);
313
314 if (hotdrop)
315 return NF_DROP;
316 else
317 return verdict;
318}
319
1da177e4
LT
320/* All zeroes == unconditional rule. */
321static inline int unconditional(const struct arpt_arp *arp)
322{
323 unsigned int i;
324
325 for (i = 0; i < sizeof(*arp)/sizeof(__u32); i++)
326 if (((__u32 *)arp)[i])
327 return 0;
328
329 return 1;
330}
331
332/* Figures out from what hook each rule can be called: returns 0 if
333 * there are loops. Puts hook bitmask in comefrom.
334 */
2e4e6a17 335static int mark_source_chains(struct xt_table_info *newinfo,
31836064 336 unsigned int valid_hooks, void *entry0)
1da177e4
LT
337{
338 unsigned int hook;
339
340 /* No recursion; use packet counter to save back ptrs (reset
341 * to 0 as we leave), and comefrom to save source hook bitmask.
342 */
343 for (hook = 0; hook < NF_ARP_NUMHOOKS; hook++) {
344 unsigned int pos = newinfo->hook_entry[hook];
345 struct arpt_entry *e
31836064 346 = (struct arpt_entry *)(entry0 + pos);
1da177e4
LT
347
348 if (!(valid_hooks & (1 << hook)))
349 continue;
350
351 /* Set initial back pointer. */
352 e->counters.pcnt = pos;
353
354 for (;;) {
355 struct arpt_standard_target *t
356 = (void *)arpt_get_target(e);
e1b4b9f3 357 int visited = e->comefrom & (1 << hook);
1da177e4
LT
358
359 if (e->comefrom & (1 << NF_ARP_NUMHOOKS)) {
360 printk("arptables: loop hook %u pos %u %08X.\n",
361 hook, pos, e->comefrom);
362 return 0;
363 }
364 e->comefrom
365 |= ((1 << hook) | (1 << NF_ARP_NUMHOOKS));
366
367 /* Unconditional return/END. */
e1b4b9f3 368 if ((e->target_offset == sizeof(struct arpt_entry)
1da177e4
LT
369 && (strcmp(t->target.u.user.name,
370 ARPT_STANDARD_TARGET) == 0)
371 && t->verdict < 0
e1b4b9f3 372 && unconditional(&e->arp)) || visited) {
1da177e4
LT
373 unsigned int oldpos, size;
374
74c9c0c1
DM
375 if (t->verdict < -NF_MAX_VERDICT - 1) {
376 duprintf("mark_source_chains: bad "
377 "negative verdict (%i)\n",
378 t->verdict);
379 return 0;
380 }
381
1da177e4
LT
382 /* Return: backtrack through the last
383 * big jump.
384 */
385 do {
386 e->comefrom ^= (1<<NF_ARP_NUMHOOKS);
387 oldpos = pos;
388 pos = e->counters.pcnt;
389 e->counters.pcnt = 0;
390
391 /* We're at the start. */
392 if (pos == oldpos)
393 goto next;
394
395 e = (struct arpt_entry *)
31836064 396 (entry0 + pos);
1da177e4
LT
397 } while (oldpos == pos + e->next_offset);
398
399 /* Move along one */
400 size = e->next_offset;
401 e = (struct arpt_entry *)
31836064 402 (entry0 + pos + size);
1da177e4
LT
403 e->counters.pcnt = pos;
404 pos += size;
405 } else {
406 int newpos = t->verdict;
407
408 if (strcmp(t->target.u.user.name,
409 ARPT_STANDARD_TARGET) == 0
410 && newpos >= 0) {
74c9c0c1
DM
411 if (newpos > newinfo->size -
412 sizeof(struct arpt_entry)) {
413 duprintf("mark_source_chains: "
414 "bad verdict (%i)\n",
415 newpos);
416 return 0;
417 }
418
1da177e4
LT
419 /* This a jump; chase it. */
420 duprintf("Jump rule %u -> %u\n",
421 pos, newpos);
422 } else {
423 /* ... this is a fallthru */
424 newpos = pos + e->next_offset;
425 }
426 e = (struct arpt_entry *)
31836064 427 (entry0 + newpos);
1da177e4
LT
428 e->counters.pcnt = pos;
429 pos = newpos;
430 }
431 }
432 next:
433 duprintf("Finished chain %u\n", hook);
434 }
435 return 1;
436}
437
fb5b6095 438static inline int check_entry(struct arpt_entry *e, const char *name)
1da177e4
LT
439{
440 struct arpt_entry_target *t;
1da177e4
LT
441
442 if (!arp_checkentry(&e->arp)) {
443 duprintf("arp_tables: arp check failed %p %s.\n", e, name);
444 return -EINVAL;
445 }
446
590bdf7f
DM
447 if (e->target_offset + sizeof(struct arpt_entry_target) > e->next_offset)
448 return -EINVAL;
449
1da177e4 450 t = arpt_get_target(e);
590bdf7f
DM
451 if (e->target_offset + t->u.target_size > e->next_offset)
452 return -EINVAL;
453
fb5b6095
PM
454 return 0;
455}
456
457static inline int check_target(struct arpt_entry *e, const char *name)
458{
459 struct arpt_entry_target *t;
460 struct arpt_target *target;
461 int ret;
462
463 t = arpt_get_target(e);
464 target = t->u.kernel.target;
465
466 ret = xt_check_target(target, NF_ARP, t->u.target_size - sizeof(*t),
467 name, e->comefrom, 0, 0);
468 if (!ret && t->u.kernel.target->checkentry
469 && !t->u.kernel.target->checkentry(name, e, target, t->data,
470 e->comefrom)) {
471 duprintf("arp_tables: check failed for `%s'.\n",
472 t->u.kernel.target->name);
473 ret = -EINVAL;
474 }
475 return ret;
476}
477
478static inline int
479find_check_entry(struct arpt_entry *e, const char *name, unsigned int size,
480 unsigned int *i)
481{
482 struct arpt_entry_target *t;
483 struct arpt_target *target;
484 int ret;
485
486 ret = check_entry(e, name);
487 if (ret)
488 return ret;
489
490 t = arpt_get_target(e);
2e4e6a17
HW
491 target = try_then_request_module(xt_find_target(NF_ARP, t->u.user.name,
492 t->u.user.revision),
6b7d31fc
HW
493 "arpt_%s", t->u.user.name);
494 if (IS_ERR(target) || !target) {
fb5b6095 495 duprintf("find_check_entry: `%s' not found\n", t->u.user.name);
6b7d31fc 496 ret = target ? PTR_ERR(target) : -ENOENT;
1da177e4
LT
497 goto out;
498 }
1da177e4 499 t->u.kernel.target = target;
1da177e4 500
fb5b6095 501 ret = check_target(e, name);
3cdc7c95
PM
502 if (ret)
503 goto err;
504
1da177e4
LT
505 (*i)++;
506 return 0;
3cdc7c95
PM
507err:
508 module_put(t->u.kernel.target->me);
1da177e4
LT
509out:
510 return ret;
511}
512
513static inline int check_entry_size_and_hooks(struct arpt_entry *e,
2e4e6a17 514 struct xt_table_info *newinfo,
1da177e4
LT
515 unsigned char *base,
516 unsigned char *limit,
517 const unsigned int *hook_entries,
518 const unsigned int *underflows,
519 unsigned int *i)
520{
521 unsigned int h;
522
523 if ((unsigned long)e % __alignof__(struct arpt_entry) != 0
524 || (unsigned char *)e + sizeof(struct arpt_entry) >= limit) {
525 duprintf("Bad offset %p\n", e);
526 return -EINVAL;
527 }
528
529 if (e->next_offset
530 < sizeof(struct arpt_entry) + sizeof(struct arpt_entry_target)) {
531 duprintf("checking: element %p size %u\n",
532 e, e->next_offset);
533 return -EINVAL;
534 }
535
536 /* Check hooks & underflows */
537 for (h = 0; h < NF_ARP_NUMHOOKS; h++) {
538 if ((unsigned char *)e - base == hook_entries[h])
539 newinfo->hook_entry[h] = hook_entries[h];
540 if ((unsigned char *)e - base == underflows[h])
541 newinfo->underflow[h] = underflows[h];
542 }
543
544 /* FIXME: underflows must be unconditional, standard verdicts
e905a9ed 545 < 0 (not ARPT_RETURN). --RR */
1da177e4
LT
546
547 /* Clear counters and comefrom */
2e4e6a17 548 e->counters = ((struct xt_counters) { 0, 0 });
1da177e4
LT
549 e->comefrom = 0;
550
551 (*i)++;
552 return 0;
553}
554
555static inline int cleanup_entry(struct arpt_entry *e, unsigned int *i)
556{
557 struct arpt_entry_target *t;
558
559 if (i && (*i)-- == 0)
560 return 1;
561
562 t = arpt_get_target(e);
563 if (t->u.kernel.target->destroy)
efa74165 564 t->u.kernel.target->destroy(t->u.kernel.target, t->data);
1da177e4
LT
565 module_put(t->u.kernel.target->me);
566 return 0;
567}
568
569/* Checks and translates the user-supplied table segment (held in
570 * newinfo).
571 */
572static int translate_table(const char *name,
573 unsigned int valid_hooks,
2e4e6a17 574 struct xt_table_info *newinfo,
31836064 575 void *entry0,
1da177e4
LT
576 unsigned int size,
577 unsigned int number,
578 const unsigned int *hook_entries,
579 const unsigned int *underflows)
580{
581 unsigned int i;
582 int ret;
583
584 newinfo->size = size;
585 newinfo->number = number;
586
587 /* Init all hooks to impossible value. */
588 for (i = 0; i < NF_ARP_NUMHOOKS; i++) {
589 newinfo->hook_entry[i] = 0xFFFFFFFF;
590 newinfo->underflow[i] = 0xFFFFFFFF;
591 }
592
593 duprintf("translate_table: size %u\n", newinfo->size);
594 i = 0;
595
596 /* Walk through entries, checking offsets. */
31836064 597 ret = ARPT_ENTRY_ITERATE(entry0, newinfo->size,
1da177e4
LT
598 check_entry_size_and_hooks,
599 newinfo,
31836064
ED
600 entry0,
601 entry0 + size,
1da177e4
LT
602 hook_entries, underflows, &i);
603 duprintf("translate_table: ARPT_ENTRY_ITERATE gives %d\n", ret);
604 if (ret != 0)
605 return ret;
606
607 if (i != number) {
608 duprintf("translate_table: %u not %u entries\n",
609 i, number);
610 return -EINVAL;
611 }
612
613 /* Check hooks all assigned */
614 for (i = 0; i < NF_ARP_NUMHOOKS; i++) {
615 /* Only hooks which are valid */
616 if (!(valid_hooks & (1 << i)))
617 continue;
618 if (newinfo->hook_entry[i] == 0xFFFFFFFF) {
619 duprintf("Invalid hook entry %u %u\n",
620 i, hook_entries[i]);
621 return -EINVAL;
622 }
623 if (newinfo->underflow[i] == 0xFFFFFFFF) {
624 duprintf("Invalid underflow %u %u\n",
625 i, underflows[i]);
626 return -EINVAL;
627 }
628 }
629
74c9c0c1
DM
630 if (!mark_source_chains(newinfo, valid_hooks, entry0)) {
631 duprintf("Looping hook\n");
632 return -ELOOP;
633 }
634
1da177e4
LT
635 /* Finally, each sanity check must pass */
636 i = 0;
31836064 637 ret = ARPT_ENTRY_ITERATE(entry0, newinfo->size,
fb5b6095 638 find_check_entry, name, size, &i);
1da177e4 639
74c9c0c1
DM
640 if (ret != 0) {
641 ARPT_ENTRY_ITERATE(entry0, newinfo->size,
642 cleanup_entry, &i);
643 return ret;
1da177e4
LT
644 }
645
646 /* And one copy for every other CPU */
6f912042 647 for_each_possible_cpu(i) {
31836064
ED
648 if (newinfo->entries[i] && newinfo->entries[i] != entry0)
649 memcpy(newinfo->entries[i], entry0, newinfo->size);
1da177e4
LT
650 }
651
652 return ret;
653}
654
1da177e4
LT
655/* Gets counters. */
656static inline int add_entry_to_counter(const struct arpt_entry *e,
2e4e6a17 657 struct xt_counters total[],
1da177e4
LT
658 unsigned int *i)
659{
660 ADD_COUNTER(total[*i], e->counters.bcnt, e->counters.pcnt);
661
662 (*i)++;
663 return 0;
664}
665
31836064 666static inline int set_entry_to_counter(const struct arpt_entry *e,
2e4e6a17 667 struct xt_counters total[],
31836064
ED
668 unsigned int *i)
669{
670 SET_COUNTER(total[*i], e->counters.bcnt, e->counters.pcnt);
671
672 (*i)++;
673 return 0;
674}
675
2e4e6a17
HW
676static void get_counters(const struct xt_table_info *t,
677 struct xt_counters counters[])
1da177e4
LT
678{
679 unsigned int cpu;
680 unsigned int i;
31836064
ED
681 unsigned int curcpu;
682
683 /* Instead of clearing (by a previous call to memset())
684 * the counters and using adds, we set the counters
685 * with data used by 'current' CPU
686 * We dont care about preemption here.
687 */
688 curcpu = raw_smp_processor_id();
689
690 i = 0;
691 ARPT_ENTRY_ITERATE(t->entries[curcpu],
692 t->size,
693 set_entry_to_counter,
694 counters,
695 &i);
1da177e4 696
6f912042 697 for_each_possible_cpu(cpu) {
31836064
ED
698 if (cpu == curcpu)
699 continue;
1da177e4 700 i = 0;
31836064 701 ARPT_ENTRY_ITERATE(t->entries[cpu],
1da177e4
LT
702 t->size,
703 add_entry_to_counter,
704 counters,
705 &i);
706 }
707}
708
27e2c26b 709static inline struct xt_counters *alloc_counters(struct arpt_table *table)
1da177e4 710{
27e2c26b 711 unsigned int countersize;
2e4e6a17
HW
712 struct xt_counters *counters;
713 struct xt_table_info *private = table->private;
1da177e4
LT
714
715 /* We need atomic snapshot of counters: rest doesn't change
716 * (other than comefrom, which userspace doesn't care
717 * about).
718 */
2e4e6a17
HW
719 countersize = sizeof(struct xt_counters) * private->number;
720 counters = vmalloc_node(countersize, numa_node_id());
1da177e4
LT
721
722 if (counters == NULL)
27e2c26b 723 return ERR_PTR(-ENOMEM);
1da177e4
LT
724
725 /* First, sum counters... */
1da177e4 726 write_lock_bh(&table->lock);
2e4e6a17 727 get_counters(private, counters);
1da177e4
LT
728 write_unlock_bh(&table->lock);
729
27e2c26b
PM
730 return counters;
731}
732
733static int copy_entries_to_user(unsigned int total_size,
734 struct arpt_table *table,
735 void __user *userptr)
736{
737 unsigned int off, num;
738 struct arpt_entry *e;
739 struct xt_counters *counters;
740 struct xt_table_info *private = table->private;
741 int ret = 0;
742 void *loc_cpu_entry;
743
744 counters = alloc_counters(table);
745 if (IS_ERR(counters))
746 return PTR_ERR(counters);
747
2e4e6a17 748 loc_cpu_entry = private->entries[raw_smp_processor_id()];
31836064
ED
749 /* ... then copy entire thing ... */
750 if (copy_to_user(userptr, loc_cpu_entry, total_size) != 0) {
1da177e4
LT
751 ret = -EFAULT;
752 goto free_counters;
753 }
754
755 /* FIXME: use iterator macros --RR */
756 /* ... then go back and fix counters and names */
757 for (off = 0, num = 0; off < total_size; off += e->next_offset, num++){
758 struct arpt_entry_target *t;
759
31836064 760 e = (struct arpt_entry *)(loc_cpu_entry + off);
1da177e4
LT
761 if (copy_to_user(userptr + off
762 + offsetof(struct arpt_entry, counters),
763 &counters[num],
764 sizeof(counters[num])) != 0) {
765 ret = -EFAULT;
766 goto free_counters;
767 }
768
769 t = arpt_get_target(e);
770 if (copy_to_user(userptr + off + e->target_offset
771 + offsetof(struct arpt_entry_target,
772 u.user.name),
773 t->u.kernel.target->name,
774 strlen(t->u.kernel.target->name)+1) != 0) {
775 ret = -EFAULT;
776 goto free_counters;
777 }
778 }
779
780 free_counters:
781 vfree(counters);
782 return ret;
783}
784
785static int get_entries(const struct arpt_get_entries *entries,
786 struct arpt_get_entries __user *uptr)
787{
788 int ret;
789 struct arpt_table *t;
790
2e4e6a17 791 t = xt_find_table_lock(NF_ARP, entries->name);
31fe4d33 792 if (t && !IS_ERR(t)) {
2e4e6a17 793 struct xt_table_info *private = t->private;
1da177e4 794 duprintf("t->private->number = %u\n",
2e4e6a17
HW
795 private->number);
796 if (entries->size == private->size)
797 ret = copy_entries_to_user(private->size,
1da177e4
LT
798 t, uptr->entrytable);
799 else {
800 duprintf("get_entries: I've got %u not %u!\n",
2e4e6a17 801 private->size, entries->size);
1da177e4
LT
802 ret = -EINVAL;
803 }
6b7d31fc 804 module_put(t->me);
2e4e6a17 805 xt_table_unlock(t);
1da177e4 806 } else
6b7d31fc 807 ret = t ? PTR_ERR(t) : -ENOENT;
1da177e4
LT
808
809 return ret;
810}
811
812static int do_replace(void __user *user, unsigned int len)
813{
814 int ret;
815 struct arpt_replace tmp;
816 struct arpt_table *t;
2e4e6a17
HW
817 struct xt_table_info *newinfo, *oldinfo;
818 struct xt_counters *counters;
31836064 819 void *loc_cpu_entry, *loc_cpu_old_entry;
1da177e4
LT
820
821 if (copy_from_user(&tmp, user, sizeof(tmp)) != 0)
822 return -EFAULT;
823
ee4bb818 824 /* overflow check */
ee4bb818
KK
825 if (tmp.num_counters >= INT_MAX / sizeof(struct xt_counters))
826 return -ENOMEM;
827
2e4e6a17 828 newinfo = xt_alloc_table_info(tmp.size);
1da177e4
LT
829 if (!newinfo)
830 return -ENOMEM;
831
31836064
ED
832 /* choose the copy that is on our node/cpu */
833 loc_cpu_entry = newinfo->entries[raw_smp_processor_id()];
834 if (copy_from_user(loc_cpu_entry, user + sizeof(tmp),
1da177e4
LT
835 tmp.size) != 0) {
836 ret = -EFAULT;
837 goto free_newinfo;
838 }
839
19763120
PM
840 counters = vmalloc_node(tmp.num_counters * sizeof(struct xt_counters),
841 numa_node_id());
1da177e4
LT
842 if (!counters) {
843 ret = -ENOMEM;
844 goto free_newinfo;
845 }
1da177e4
LT
846
847 ret = translate_table(tmp.name, tmp.valid_hooks,
31836064 848 newinfo, loc_cpu_entry, tmp.size, tmp.num_entries,
1da177e4
LT
849 tmp.hook_entry, tmp.underflow);
850 if (ret != 0)
851 goto free_newinfo_counters;
852
853 duprintf("arp_tables: Translated table\n");
854
2e4e6a17 855 t = try_then_request_module(xt_find_table_lock(NF_ARP, tmp.name),
6b7d31fc
HW
856 "arptable_%s", tmp.name);
857 if (!t || IS_ERR(t)) {
858 ret = t ? PTR_ERR(t) : -ENOENT;
1da177e4 859 goto free_newinfo_counters_untrans;
6b7d31fc 860 }
1da177e4
LT
861
862 /* You lied! */
863 if (tmp.valid_hooks != t->valid_hooks) {
864 duprintf("Valid hook crap: %08X vs %08X\n",
865 tmp.valid_hooks, t->valid_hooks);
866 ret = -EINVAL;
6b7d31fc 867 goto put_module;
1da177e4
LT
868 }
869
2e4e6a17 870 oldinfo = xt_replace_table(t, tmp.num_counters, newinfo, &ret);
1da177e4
LT
871 if (!oldinfo)
872 goto put_module;
873
874 /* Update module usage count based on number of rules */
875 duprintf("do_replace: oldnum=%u, initnum=%u, newnum=%u\n",
876 oldinfo->number, oldinfo->initial_entries, newinfo->number);
e905a9ed
YH
877 if ((oldinfo->number > oldinfo->initial_entries) ||
878 (newinfo->number <= oldinfo->initial_entries))
1da177e4
LT
879 module_put(t->me);
880 if ((oldinfo->number > oldinfo->initial_entries) &&
881 (newinfo->number <= oldinfo->initial_entries))
882 module_put(t->me);
883
884 /* Get the old counters. */
885 get_counters(oldinfo, counters);
886 /* Decrease module usage counts and free resource */
31836064
ED
887 loc_cpu_old_entry = oldinfo->entries[raw_smp_processor_id()];
888 ARPT_ENTRY_ITERATE(loc_cpu_old_entry, oldinfo->size, cleanup_entry,NULL);
889
2e4e6a17 890 xt_free_table_info(oldinfo);
1da177e4 891 if (copy_to_user(tmp.counters, counters,
2e4e6a17 892 sizeof(struct xt_counters) * tmp.num_counters) != 0)
1da177e4
LT
893 ret = -EFAULT;
894 vfree(counters);
2e4e6a17 895 xt_table_unlock(t);
1da177e4
LT
896 return ret;
897
898 put_module:
899 module_put(t->me);
2e4e6a17 900 xt_table_unlock(t);
1da177e4 901 free_newinfo_counters_untrans:
31836064 902 ARPT_ENTRY_ITERATE(loc_cpu_entry, newinfo->size, cleanup_entry, NULL);
1da177e4
LT
903 free_newinfo_counters:
904 vfree(counters);
905 free_newinfo:
2e4e6a17 906 xt_free_table_info(newinfo);
1da177e4
LT
907 return ret;
908}
909
910/* We're lazy, and add to the first CPU; overflow works its fey magic
911 * and everything is OK.
912 */
913static inline int add_counter_to_entry(struct arpt_entry *e,
2e4e6a17 914 const struct xt_counters addme[],
1da177e4
LT
915 unsigned int *i)
916{
917
918 ADD_COUNTER(e->counters, addme[*i].bcnt, addme[*i].pcnt);
919
920 (*i)++;
921 return 0;
922}
923
924static int do_add_counters(void __user *user, unsigned int len)
925{
926 unsigned int i;
2e4e6a17 927 struct xt_counters_info tmp, *paddc;
1da177e4 928 struct arpt_table *t;
2e4e6a17 929 struct xt_table_info *private;
6b7d31fc 930 int ret = 0;
31836064 931 void *loc_cpu_entry;
1da177e4
LT
932
933 if (copy_from_user(&tmp, user, sizeof(tmp)) != 0)
934 return -EFAULT;
935
2e4e6a17 936 if (len != sizeof(tmp) + tmp.num_counters*sizeof(struct xt_counters))
1da177e4
LT
937 return -EINVAL;
938
19763120 939 paddc = vmalloc_node(len, numa_node_id());
1da177e4
LT
940 if (!paddc)
941 return -ENOMEM;
942
943 if (copy_from_user(paddc, user, len) != 0) {
944 ret = -EFAULT;
945 goto free;
946 }
947
2e4e6a17 948 t = xt_find_table_lock(NF_ARP, tmp.name);
6b7d31fc
HW
949 if (!t || IS_ERR(t)) {
950 ret = t ? PTR_ERR(t) : -ENOENT;
1da177e4 951 goto free;
6b7d31fc 952 }
1da177e4
LT
953
954 write_lock_bh(&t->lock);
2e4e6a17 955 private = t->private;
2c8ac66b 956 if (private->number != tmp.num_counters) {
1da177e4
LT
957 ret = -EINVAL;
958 goto unlock_up_free;
959 }
960
961 i = 0;
31836064 962 /* Choose the copy that is on our node */
2e4e6a17 963 loc_cpu_entry = private->entries[smp_processor_id()];
31836064 964 ARPT_ENTRY_ITERATE(loc_cpu_entry,
2e4e6a17 965 private->size,
1da177e4
LT
966 add_counter_to_entry,
967 paddc->counters,
968 &i);
969 unlock_up_free:
970 write_unlock_bh(&t->lock);
2e4e6a17 971 xt_table_unlock(t);
6b7d31fc 972 module_put(t->me);
1da177e4
LT
973 free:
974 vfree(paddc);
975
976 return ret;
977}
978
979static int do_arpt_set_ctl(struct sock *sk, int cmd, void __user *user, unsigned int len)
980{
981 int ret;
982
983 if (!capable(CAP_NET_ADMIN))
984 return -EPERM;
985
986 switch (cmd) {
987 case ARPT_SO_SET_REPLACE:
988 ret = do_replace(user, len);
989 break;
990
991 case ARPT_SO_SET_ADD_COUNTERS:
992 ret = do_add_counters(user, len);
993 break;
994
995 default:
996 duprintf("do_arpt_set_ctl: unknown request %i\n", cmd);
997 ret = -EINVAL;
998 }
999
1000 return ret;
1001}
1002
1003static int do_arpt_get_ctl(struct sock *sk, int cmd, void __user *user, int *len)
1004{
1005 int ret;
1006
1007 if (!capable(CAP_NET_ADMIN))
1008 return -EPERM;
1009
1010 switch (cmd) {
1011 case ARPT_SO_GET_INFO: {
1012 char name[ARPT_TABLE_MAXNAMELEN];
1013 struct arpt_table *t;
1014
1015 if (*len != sizeof(struct arpt_getinfo)) {
1016 duprintf("length %u != %Zu\n", *len,
1017 sizeof(struct arpt_getinfo));
1018 ret = -EINVAL;
1019 break;
1020 }
1021
1022 if (copy_from_user(name, user, sizeof(name)) != 0) {
1023 ret = -EFAULT;
1024 break;
1025 }
1026 name[ARPT_TABLE_MAXNAMELEN-1] = '\0';
6b7d31fc 1027
2e4e6a17 1028 t = try_then_request_module(xt_find_table_lock(NF_ARP, name),
6b7d31fc
HW
1029 "arptable_%s", name);
1030 if (t && !IS_ERR(t)) {
1da177e4 1031 struct arpt_getinfo info;
2e4e6a17 1032 struct xt_table_info *private = t->private;
1da177e4
LT
1033
1034 info.valid_hooks = t->valid_hooks;
2e4e6a17 1035 memcpy(info.hook_entry, private->hook_entry,
1da177e4 1036 sizeof(info.hook_entry));
2e4e6a17 1037 memcpy(info.underflow, private->underflow,
1da177e4 1038 sizeof(info.underflow));
2e4e6a17
HW
1039 info.num_entries = private->number;
1040 info.size = private->size;
1da177e4
LT
1041 strcpy(info.name, name);
1042
1043 if (copy_to_user(user, &info, *len) != 0)
1044 ret = -EFAULT;
1045 else
1046 ret = 0;
2e4e6a17 1047 xt_table_unlock(t);
6b7d31fc
HW
1048 module_put(t->me);
1049 } else
1050 ret = t ? PTR_ERR(t) : -ENOENT;
1da177e4
LT
1051 }
1052 break;
1053
1054 case ARPT_SO_GET_ENTRIES: {
1055 struct arpt_get_entries get;
1056
1057 if (*len < sizeof(get)) {
1058 duprintf("get_entries: %u < %Zu\n", *len, sizeof(get));
1059 ret = -EINVAL;
1060 } else if (copy_from_user(&get, user, sizeof(get)) != 0) {
1061 ret = -EFAULT;
1062 } else if (*len != sizeof(struct arpt_get_entries) + get.size) {
1063 duprintf("get_entries: %u != %Zu\n", *len,
1064 sizeof(struct arpt_get_entries) + get.size);
1065 ret = -EINVAL;
1066 } else
1067 ret = get_entries(&get, user);
1068 break;
1069 }
1070
6b7d31fc 1071 case ARPT_SO_GET_REVISION_TARGET: {
2e4e6a17 1072 struct xt_get_revision rev;
6b7d31fc
HW
1073
1074 if (*len != sizeof(rev)) {
1075 ret = -EINVAL;
1076 break;
1077 }
1078 if (copy_from_user(&rev, user, sizeof(rev)) != 0) {
1079 ret = -EFAULT;
1080 break;
1081 }
1082
2e4e6a17
HW
1083 try_then_request_module(xt_find_revision(NF_ARP, rev.name,
1084 rev.revision, 1, &ret),
6b7d31fc
HW
1085 "arpt_%s", rev.name);
1086 break;
1087 }
1088
1da177e4
LT
1089 default:
1090 duprintf("do_arpt_get_ctl: unknown request %i\n", cmd);
1091 ret = -EINVAL;
1092 }
1093
1094 return ret;
1095}
1096
1da177e4
LT
1097int arpt_register_table(struct arpt_table *table,
1098 const struct arpt_replace *repl)
1099{
1100 int ret;
2e4e6a17 1101 struct xt_table_info *newinfo;
259d4e41 1102 struct xt_table_info bootstrap
1da177e4 1103 = { 0, 0, 0, { 0 }, { 0 }, { } };
31836064 1104 void *loc_cpu_entry;
1da177e4 1105
2e4e6a17 1106 newinfo = xt_alloc_table_info(repl->size);
1da177e4
LT
1107 if (!newinfo) {
1108 ret = -ENOMEM;
1109 return ret;
1110 }
31836064
ED
1111
1112 /* choose the copy on our node/cpu */
1113 loc_cpu_entry = newinfo->entries[raw_smp_processor_id()];
1114 memcpy(loc_cpu_entry, repl->entries, repl->size);
1da177e4
LT
1115
1116 ret = translate_table(table->name, table->valid_hooks,
31836064 1117 newinfo, loc_cpu_entry, repl->size,
1da177e4
LT
1118 repl->num_entries,
1119 repl->hook_entry,
1120 repl->underflow);
2e4e6a17 1121
1da177e4
LT
1122 duprintf("arpt_register_table: translate table gives %d\n", ret);
1123 if (ret != 0) {
2e4e6a17 1124 xt_free_table_info(newinfo);
1da177e4
LT
1125 return ret;
1126 }
1127
da298d3a
PM
1128 ret = xt_register_table(table, &bootstrap, newinfo);
1129 if (ret != 0) {
2e4e6a17 1130 xt_free_table_info(newinfo);
1da177e4
LT
1131 return ret;
1132 }
1133
2e4e6a17 1134 return 0;
1da177e4
LT
1135}
1136
1137void arpt_unregister_table(struct arpt_table *table)
1138{
2e4e6a17 1139 struct xt_table_info *private;
31836064
ED
1140 void *loc_cpu_entry;
1141
2e4e6a17 1142 private = xt_unregister_table(table);
1da177e4
LT
1143
1144 /* Decrease module usage counts and free resources */
2e4e6a17
HW
1145 loc_cpu_entry = private->entries[raw_smp_processor_id()];
1146 ARPT_ENTRY_ITERATE(loc_cpu_entry, private->size,
1da177e4 1147 cleanup_entry, NULL);
2e4e6a17 1148 xt_free_table_info(private);
1da177e4
LT
1149}
1150
1151/* The built-in targets: standard (NULL) and error. */
9f15c530 1152static struct arpt_target arpt_standard_target __read_mostly = {
1da177e4 1153 .name = ARPT_STANDARD_TARGET,
aa83c1ab 1154 .targetsize = sizeof(int),
a45049c5 1155 .family = NF_ARP,
1da177e4
LT
1156};
1157
9f15c530 1158static struct arpt_target arpt_error_target __read_mostly = {
1da177e4
LT
1159 .name = ARPT_ERROR_TARGET,
1160 .target = arpt_error,
aa83c1ab 1161 .targetsize = ARPT_FUNCTION_MAXNAMELEN,
a45049c5 1162 .family = NF_ARP,
1da177e4
LT
1163};
1164
1165static struct nf_sockopt_ops arpt_sockopts = {
1166 .pf = PF_INET,
1167 .set_optmin = ARPT_BASE_CTL,
1168 .set_optmax = ARPT_SO_SET_MAX+1,
1169 .set = do_arpt_set_ctl,
1170 .get_optmin = ARPT_BASE_CTL,
1171 .get_optmax = ARPT_SO_GET_MAX+1,
1172 .get = do_arpt_get_ctl,
16fcec35 1173 .owner = THIS_MODULE,
1da177e4
LT
1174};
1175
65b4b4e8 1176static int __init arp_tables_init(void)
1da177e4
LT
1177{
1178 int ret;
1179
0eff66e6
PM
1180 ret = xt_proto_init(NF_ARP);
1181 if (ret < 0)
1182 goto err1;
2e4e6a17 1183
1da177e4 1184 /* Noone else will be downing sem now, so we won't sleep */
0eff66e6
PM
1185 ret = xt_register_target(&arpt_standard_target);
1186 if (ret < 0)
1187 goto err2;
1188 ret = xt_register_target(&arpt_error_target);
1189 if (ret < 0)
1190 goto err3;
1da177e4
LT
1191
1192 /* Register setsockopt */
1193 ret = nf_register_sockopt(&arpt_sockopts);
0eff66e6
PM
1194 if (ret < 0)
1195 goto err4;
1da177e4 1196
a887c1c1 1197 printk(KERN_INFO "arp_tables: (C) 2002 David S. Miller\n");
1da177e4 1198 return 0;
0eff66e6
PM
1199
1200err4:
1201 xt_unregister_target(&arpt_error_target);
1202err3:
1203 xt_unregister_target(&arpt_standard_target);
1204err2:
1205 xt_proto_fini(NF_ARP);
1206err1:
1207 return ret;
1da177e4
LT
1208}
1209
65b4b4e8 1210static void __exit arp_tables_fini(void)
1da177e4
LT
1211{
1212 nf_unregister_sockopt(&arpt_sockopts);
f603b6ec
PM
1213 xt_unregister_target(&arpt_error_target);
1214 xt_unregister_target(&arpt_standard_target);
2e4e6a17 1215 xt_proto_fini(NF_ARP);
1da177e4
LT
1216}
1217
1218EXPORT_SYMBOL(arpt_register_table);
1219EXPORT_SYMBOL(arpt_unregister_table);
1220EXPORT_SYMBOL(arpt_do_table);
1da177e4 1221
65b4b4e8
AM
1222module_init(arp_tables_init);
1223module_exit(arp_tables_fini);