kbuild: add headerdep used to detect inclusion cycles in header files
[GitHub/mt8127/android_kernel_alcatel_ttab.git] / include / linux / netpoll.h
CommitLineData
1da177e4
LT
1/*
2 * Common code for low-level network console, dump, and debugger code
3 *
4 * Derived from netconsole, kgdb-over-ethernet, and netdump patches
5 */
6
7#ifndef _LINUX_NETPOLL_H
8#define _LINUX_NETPOLL_H
9
10#include <linux/netdevice.h>
11#include <linux/interrupt.h>
53fb95d3 12#include <linux/rcupdate.h>
1da177e4
LT
13#include <linux/list.h>
14
1da177e4
LT
15struct netpoll {
16 struct net_device *dev;
bf6bce71
SH
17 char dev_name[IFNAMSIZ];
18 const char *name;
1da177e4 19 void (*rx_hook)(struct netpoll *, int, char *, int);
5de4a473 20
1da177e4
LT
21 u32 local_ip, remote_ip;
22 u16 local_port, remote_port;
09538641 23 u8 remote_mac[ETH_ALEN];
115c1d6e
JM
24};
25
26struct netpoll_info {
93ec2c72 27 atomic_t refcnt;
d9452e9f 28 int rx_flags;
fbeec2e1
JM
29 spinlock_t rx_lock;
30 struct netpoll *rx_np; /* netpoll that registered an rx_hook */
068c6e98 31 struct sk_buff_head arp_tx; /* list of arp requests to reply to */
b6cd27ed 32 struct sk_buff_head txq;
6d5aefb8 33 struct delayed_work tx_work;
1da177e4
LT
34};
35
36void netpoll_poll(struct netpoll *np);
37void netpoll_send_udp(struct netpoll *np, const char *msg, int len);
0bcc1816 38void netpoll_print_options(struct netpoll *np);
1da177e4
LT
39int netpoll_parse_options(struct netpoll *np, char *opt);
40int netpoll_setup(struct netpoll *np);
41int netpoll_trap(void);
42void netpoll_set_trap(int trap);
43void netpoll_cleanup(struct netpoll *np);
44int __netpoll_rx(struct sk_buff *skb);
5de4a473 45
1da177e4
LT
46
47#ifdef CONFIG_NETPOLL
48static inline int netpoll_rx(struct sk_buff *skb)
49{
115c1d6e 50 struct netpoll_info *npinfo = skb->dev->npinfo;
fbeec2e1
JM
51 unsigned long flags;
52 int ret = 0;
115c1d6e 53
d9452e9f 54 if (!npinfo || (!npinfo->rx_np && !npinfo->rx_flags))
115c1d6e
JM
55 return 0;
56
fbeec2e1 57 spin_lock_irqsave(&npinfo->rx_lock, flags);
d9452e9f
DM
58 /* check rx_flags again with the lock held */
59 if (npinfo->rx_flags && __netpoll_rx(skb))
fbeec2e1
JM
60 ret = 1;
61 spin_unlock_irqrestore(&npinfo->rx_lock, flags);
62
63 return ret;
1da177e4
LT
64}
65
bea3348e 66static inline int netpoll_receive_skb(struct sk_buff *skb)
1da177e4 67{
bea3348e
SH
68 if (!list_empty(&skb->dev->napi_list))
69 return netpoll_rx(skb);
70 return 0;
71}
72
73static inline void *netpoll_poll_lock(struct napi_struct *napi)
74{
75 struct net_device *dev = napi->dev;
76
53fb95d3 77 rcu_read_lock(); /* deal with race on ->npinfo */
bea3348e
SH
78 if (dev && dev->npinfo) {
79 spin_lock(&napi->poll_lock);
80 napi->poll_owner = smp_processor_id();
81 return napi;
1da177e4 82 }
53fb95d3 83 return NULL;
1da177e4
LT
84}
85
53fb95d3 86static inline void netpoll_poll_unlock(void *have)
1da177e4 87{
bea3348e 88 struct napi_struct *napi = have;
53fb95d3 89
bea3348e
SH
90 if (napi) {
91 napi->poll_owner = -1;
92 spin_unlock(&napi->poll_lock);
1da177e4 93 }
53fb95d3 94 rcu_read_unlock();
1da177e4
LT
95}
96
bea3348e
SH
97static inline void netpoll_netdev_init(struct net_device *dev)
98{
99 INIT_LIST_HEAD(&dev->napi_list);
100}
101
1da177e4 102#else
bea3348e
SH
103static inline int netpoll_rx(struct sk_buff *skb)
104{
105 return 0;
106}
107static inline int netpoll_receive_skb(struct sk_buff *skb)
108{
109 return 0;
110}
111static inline void *netpoll_poll_lock(struct napi_struct *napi)
112{
113 return NULL;
114}
115static inline void netpoll_poll_unlock(void *have)
116{
117}
118static inline void netpoll_netdev_init(struct net_device *dev)
119{
120}
1da177e4
LT
121#endif
122
123#endif