Commit | Line | Data |
---|---|---|
1da177e4 | 1 | /* |
cd1ae0e4 | 2 | * Copyright (C) 2001 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com) |
f28169d2 | 3 | * Copyright (C) 2001 Lennert Buytenhek (buytenh@gnu.org) and |
1da177e4 LT |
4 | * James Leu (jleu@mindspring.net). |
5 | * Copyright (C) 2001 by various other people who didn't put their name here. | |
6 | * Licensed under the GPL. | |
7 | */ | |
8 | ||
cd1ae0e4 JD |
9 | #include <linux/bootmem.h> |
10 | #include <linux/etherdevice.h> | |
11 | #include <linux/ethtool.h> | |
12 | #include <linux/inetdevice.h> | |
13 | #include <linux/init.h> | |
14 | #include <linux/list.h> | |
15 | #include <linux/netdevice.h> | |
16 | #include <linux/platform_device.h> | |
17 | #include <linux/rtnetlink.h> | |
18 | #include <linux/skbuff.h> | |
19 | #include <linux/spinlock.h> | |
1da177e4 | 20 | #include "init.h" |
1da177e4 | 21 | #include "irq_kern.h" |
cd1ae0e4 JD |
22 | #include "irq_user.h" |
23 | #include "mconsole_kern.h" | |
24 | #include "net_kern.h" | |
25 | #include "net_user.h" | |
1da177e4 | 26 | |
bf61f50d PBG |
27 | static inline void set_ether_mac(struct net_device *dev, unsigned char *addr) |
28 | { | |
29 | memcpy(dev->dev_addr, addr, ETH_ALEN); | |
30 | } | |
31 | ||
1da177e4 LT |
32 | #define DRIVER_NAME "uml-netdev" |
33 | ||
34 | static DEFINE_SPINLOCK(opened_lock); | |
9010772c | 35 | static LIST_HEAD(opened); |
1da177e4 | 36 | |
605c1e57 JD |
37 | /* |
38 | * The drop_skb is used when we can't allocate an skb. The | |
39 | * packet is read into drop_skb in order to get the data off the | |
40 | * connection to the host. | |
41 | * It is reallocated whenever a maximum packet size is seen which is | |
42 | * larger than any seen before. update_drop_skb is called from | |
43 | * eth_configure when a new interface is added. | |
44 | */ | |
45 | static DEFINE_SPINLOCK(drop_lock); | |
46 | static struct sk_buff *drop_skb; | |
47 | static int drop_max; | |
48 | ||
49 | static int update_drop_skb(int max) | |
50 | { | |
51 | struct sk_buff *new; | |
52 | unsigned long flags; | |
53 | int err = 0; | |
54 | ||
55 | spin_lock_irqsave(&drop_lock, flags); | |
56 | ||
57 | if (max <= drop_max) | |
58 | goto out; | |
59 | ||
60 | err = -ENOMEM; | |
61 | new = dev_alloc_skb(max); | |
62 | if (new == NULL) | |
63 | goto out; | |
64 | ||
65 | skb_put(new, max); | |
66 | ||
67 | kfree_skb(drop_skb); | |
68 | drop_skb = new; | |
69 | drop_max = max; | |
70 | err = 0; | |
71 | out: | |
72 | spin_unlock_irqrestore(&drop_lock, flags); | |
73 | ||
74 | return err; | |
75 | } | |
76 | ||
1da177e4 LT |
77 | static int uml_net_rx(struct net_device *dev) |
78 | { | |
79 | struct uml_net_private *lp = dev->priv; | |
80 | int pkt_len; | |
81 | struct sk_buff *skb; | |
82 | ||
83 | /* If we can't allocate memory, try again next round. */ | |
b53f35a8 | 84 | skb = dev_alloc_skb(lp->max_packet); |
1da177e4 | 85 | if (skb == NULL) { |
605c1e57 JD |
86 | drop_skb->dev = dev; |
87 | /* Read a packet into drop_skb and don't do anything with it. */ | |
88 | (*lp->read)(lp->fd, drop_skb, lp); | |
1da177e4 LT |
89 | lp->stats.rx_dropped++; |
90 | return 0; | |
91 | } | |
92 | ||
93 | skb->dev = dev; | |
b53f35a8 | 94 | skb_put(skb, lp->max_packet); |
459a98ed | 95 | skb_reset_mac_header(skb); |
b53f35a8 | 96 | pkt_len = (*lp->read)(lp->fd, skb, lp); |
1da177e4 LT |
97 | |
98 | if (pkt_len > 0) { | |
99 | skb_trim(skb, pkt_len); | |
100 | skb->protocol = (*lp->protocol)(skb); | |
1da177e4 LT |
101 | |
102 | lp->stats.rx_bytes += skb->len; | |
103 | lp->stats.rx_packets++; | |
505a41d4 | 104 | netif_rx(skb); |
1da177e4 LT |
105 | return pkt_len; |
106 | } | |
107 | ||
108 | kfree_skb(skb); | |
109 | return pkt_len; | |
110 | } | |
111 | ||
eff3b634 | 112 | static void uml_dev_close(struct work_struct *work) |
71c8d4c3 | 113 | { |
eff3b634 PZ |
114 | struct uml_net_private *lp = |
115 | container_of(work, struct uml_net_private, work); | |
116 | dev_close(lp->dev); | |
71c8d4c3 PBG |
117 | } |
118 | ||
7bea96fd | 119 | irqreturn_t uml_net_interrupt(int irq, void *dev_id) |
1da177e4 LT |
120 | { |
121 | struct net_device *dev = dev_id; | |
122 | struct uml_net_private *lp = dev->priv; | |
123 | int err; | |
124 | ||
cd1ae0e4 JD |
125 | if (!netif_running(dev)) |
126 | return IRQ_NONE; | |
1da177e4 LT |
127 | |
128 | spin_lock(&lp->lock); | |
cd1ae0e4 JD |
129 | while ((err = uml_net_rx(dev)) > 0) ; |
130 | if (err < 0) { | |
f28169d2 JD |
131 | printk(KERN_ERR |
132 | "Device '%s' read returned %d, shutting it down\n", | |
1da177e4 | 133 | dev->name, err); |
71c8d4c3 PBG |
134 | /* dev_close can't be called in interrupt context, and takes |
135 | * again lp->lock. | |
136 | * And dev_close() can be safely called multiple times on the | |
137 | * same device, since it tests for (dev->flags & IFF_UP). So | |
eff3b634 PZ |
138 | * there's no harm in delaying the device shutdown. |
139 | * Furthermore, the workqueue will not re-enqueue an already | |
140 | * enqueued work item. */ | |
141 | schedule_work(&lp->work); | |
1da177e4 LT |
142 | goto out; |
143 | } | |
144 | reactivate_fd(lp->fd, UM_ETH_IRQ); | |
145 | ||
71c8d4c3 | 146 | out: |
1da177e4 | 147 | spin_unlock(&lp->lock); |
4ea21cd9 | 148 | return IRQ_HANDLED; |
1da177e4 LT |
149 | } |
150 | ||
151 | static int uml_net_open(struct net_device *dev) | |
152 | { | |
153 | struct uml_net_private *lp = dev->priv; | |
1da177e4 LT |
154 | int err; |
155 | ||
cd1ae0e4 | 156 | if (lp->fd >= 0) { |
1da177e4 LT |
157 | err = -ENXIO; |
158 | goto out; | |
159 | } | |
160 | ||
1da177e4 | 161 | lp->fd = (*lp->open)(&lp->user); |
cd1ae0e4 | 162 | if (lp->fd < 0) { |
1da177e4 LT |
163 | err = lp->fd; |
164 | goto out; | |
165 | } | |
166 | ||
167 | err = um_request_irq(dev->irq, lp->fd, IRQ_READ, uml_net_interrupt, | |
bd6aa650 | 168 | IRQF_DISABLED | IRQF_SHARED, dev->name, dev); |
cd1ae0e4 | 169 | if (err != 0) { |
1da177e4 | 170 | printk(KERN_ERR "uml_net_open: failed to get irq(%d)\n", err); |
1da177e4 | 171 | err = -ENETUNREACH; |
14d9ead0 | 172 | goto out_close; |
1da177e4 LT |
173 | } |
174 | ||
175 | lp->tl.data = (unsigned long) &lp->user; | |
176 | netif_start_queue(dev); | |
177 | ||
178 | /* clear buffer - it can happen that the host side of the interface | |
179 | * is full when we get here. In this case, new data is never queued, | |
180 | * SIGIOs never arrive, and the net never works. | |
181 | */ | |
cd1ae0e4 | 182 | while ((err = uml_net_rx(dev)) > 0) ; |
1da177e4 | 183 | |
14d9ead0 JD |
184 | spin_lock(&opened_lock); |
185 | list_add(&lp->list, &opened); | |
186 | spin_unlock(&opened_lock); | |
187 | ||
188 | return 0; | |
189 | out_close: | |
cd1ae0e4 | 190 | if (lp->close != NULL) (*lp->close)(lp->fd, &lp->user); |
14d9ead0 JD |
191 | lp->fd = -1; |
192 | out: | |
14d9ead0 | 193 | return err; |
1da177e4 LT |
194 | } |
195 | ||
196 | static int uml_net_close(struct net_device *dev) | |
197 | { | |
198 | struct uml_net_private *lp = dev->priv; | |
f28169d2 | 199 | |
1da177e4 | 200 | netif_stop_queue(dev); |
1da177e4 | 201 | |
1da177e4 | 202 | free_irq(dev->irq, dev); |
cd1ae0e4 | 203 | if (lp->close != NULL) |
1da177e4 LT |
204 | (*lp->close)(lp->fd, &lp->user); |
205 | lp->fd = -1; | |
206 | ||
14d9ead0 JD |
207 | spin_lock(&opened_lock); |
208 | list_del(&lp->list); | |
209 | spin_unlock(&opened_lock); | |
210 | ||
1da177e4 LT |
211 | return 0; |
212 | } | |
213 | ||
214 | static int uml_net_start_xmit(struct sk_buff *skb, struct net_device *dev) | |
215 | { | |
216 | struct uml_net_private *lp = dev->priv; | |
217 | unsigned long flags; | |
218 | int len; | |
219 | ||
220 | netif_stop_queue(dev); | |
221 | ||
222 | spin_lock_irqsave(&lp->lock, flags); | |
223 | ||
b53f35a8 | 224 | len = (*lp->write)(lp->fd, skb, lp); |
1da177e4 | 225 | |
cd1ae0e4 | 226 | if (len == skb->len) { |
1da177e4 LT |
227 | lp->stats.tx_packets++; |
228 | lp->stats.tx_bytes += skb->len; | |
229 | dev->trans_start = jiffies; | |
230 | netif_start_queue(dev); | |
231 | ||
232 | /* this is normally done in the interrupt when tx finishes */ | |
233 | netif_wake_queue(dev); | |
f28169d2 | 234 | } |
cd1ae0e4 | 235 | else if (len == 0) { |
1da177e4 LT |
236 | netif_start_queue(dev); |
237 | lp->stats.tx_dropped++; | |
238 | } | |
239 | else { | |
240 | netif_start_queue(dev); | |
241 | printk(KERN_ERR "uml_net_start_xmit: failed(%d)\n", len); | |
242 | } | |
243 | ||
244 | spin_unlock_irqrestore(&lp->lock, flags); | |
245 | ||
246 | dev_kfree_skb(skb); | |
247 | ||
248 | return 0; | |
249 | } | |
250 | ||
251 | static struct net_device_stats *uml_net_get_stats(struct net_device *dev) | |
252 | { | |
253 | struct uml_net_private *lp = dev->priv; | |
254 | return &lp->stats; | |
255 | } | |
256 | ||
257 | static void uml_net_set_multicast_list(struct net_device *dev) | |
258 | { | |
cd1ae0e4 JD |
259 | if (dev->flags & IFF_PROMISC) |
260 | return; | |
261 | else if (dev->mc_count) | |
262 | dev->flags |= IFF_ALLMULTI; | |
1da177e4 LT |
263 | else dev->flags &= ~IFF_ALLMULTI; |
264 | } | |
265 | ||
266 | static void uml_net_tx_timeout(struct net_device *dev) | |
267 | { | |
268 | dev->trans_start = jiffies; | |
269 | netif_wake_queue(dev); | |
270 | } | |
271 | ||
272 | static int uml_net_set_mac(struct net_device *dev, void *addr) | |
273 | { | |
274 | struct uml_net_private *lp = dev->priv; | |
275 | struct sockaddr *hwaddr = addr; | |
276 | ||
48af05ed | 277 | spin_lock_irq(&lp->lock); |
bf61f50d | 278 | set_ether_mac(dev, hwaddr->sa_data); |
48af05ed | 279 | spin_unlock_irq(&lp->lock); |
1da177e4 | 280 | |
4ea21cd9 | 281 | return 0; |
1da177e4 LT |
282 | } |
283 | ||
284 | static int uml_net_change_mtu(struct net_device *dev, int new_mtu) | |
285 | { | |
1da177e4 LT |
286 | dev->mtu = new_mtu; |
287 | ||
b53f35a8 | 288 | return 0; |
1da177e4 LT |
289 | } |
290 | ||
6d387484 CH |
291 | static void uml_net_get_drvinfo(struct net_device *dev, |
292 | struct ethtool_drvinfo *info) | |
1da177e4 | 293 | { |
6d387484 CH |
294 | strcpy(info->driver, DRIVER_NAME); |
295 | strcpy(info->version, "42"); | |
1da177e4 LT |
296 | } |
297 | ||
6d387484 CH |
298 | static struct ethtool_ops uml_net_ethtool_ops = { |
299 | .get_drvinfo = uml_net_get_drvinfo, | |
300 | .get_link = ethtool_op_get_link, | |
301 | }; | |
302 | ||
1da177e4 LT |
303 | void uml_net_user_timer_expire(unsigned long _conn) |
304 | { | |
305 | #ifdef undef | |
306 | struct connection *conn = (struct connection *)_conn; | |
307 | ||
308 | dprintk(KERN_INFO "uml_net_user_timer_expire [%p]\n", conn); | |
309 | do_connect(conn); | |
310 | #endif | |
311 | } | |
312 | ||
e024715f | 313 | static void setup_etheraddr(char *str, unsigned char *addr, char *name) |
b10aeeef JD |
314 | { |
315 | char *end; | |
316 | int i; | |
317 | ||
cd1ae0e4 | 318 | if (str == NULL) |
b10aeeef JD |
319 | goto random; |
320 | ||
2278c5ac | 321 | for (i = 0; i < 6; i++) { |
b10aeeef | 322 | addr[i] = simple_strtoul(str, &end, 16); |
cd1ae0e4 JD |
323 | if ((end == str) || |
324 | ((*end != ':') && (*end != ',') && (*end != '\0'))) { | |
b10aeeef JD |
325 | printk(KERN_ERR |
326 | "setup_etheraddr: failed to parse '%s' " | |
327 | "as an ethernet address\n", str); | |
328 | goto random; | |
329 | } | |
330 | str = end + 1; | |
331 | } | |
e024715f | 332 | if (is_multicast_ether_addr(addr)) { |
b10aeeef | 333 | printk(KERN_ERR |
e024715f | 334 | "Attempt to assign a multicast ethernet address to a " |
b10aeeef JD |
335 | "device disallowed\n"); |
336 | goto random; | |
337 | } | |
e024715f PBG |
338 | if (!is_valid_ether_addr(addr)) { |
339 | printk(KERN_ERR | |
340 | "Attempt to assign an invalid ethernet address to a " | |
341 | "device disallowed\n"); | |
342 | goto random; | |
343 | } | |
344 | if (!is_local_ether_addr(addr)) { | |
345 | printk(KERN_WARNING | |
2278c5ac | 346 | "Warning: Assigning a globally valid ethernet " |
7d98230a | 347 | "address to a device\n"); |
2278c5ac JD |
348 | printk(KERN_WARNING "You should set the 2nd rightmost bit in " |
349 | "the first byte of the MAC,\n"); | |
7d98230a JD |
350 | printk(KERN_WARNING "i.e. %02x:%02x:%02x:%02x:%02x:%02x\n", |
351 | addr[0] | 0x02, addr[1], addr[2], addr[3], addr[4], | |
352 | addr[5]); | |
e024715f | 353 | } |
b10aeeef JD |
354 | return; |
355 | ||
356 | random: | |
e024715f PBG |
357 | printk(KERN_INFO |
358 | "Choosing a random ethernet address for device %s\n", name); | |
d6c64102 | 359 | random_ether_addr(addr); |
b10aeeef JD |
360 | } |
361 | ||
1da177e4 | 362 | static DEFINE_SPINLOCK(devices_lock); |
9010772c | 363 | static LIST_HEAD(devices); |
1da177e4 | 364 | |
3ae5eaec RK |
365 | static struct platform_driver uml_net_driver = { |
366 | .driver = { | |
367 | .name = DRIVER_NAME, | |
368 | }, | |
1da177e4 | 369 | }; |
1da177e4 | 370 | |
2e3f5251 JD |
371 | static void net_device_release(struct device *dev) |
372 | { | |
373 | struct uml_net *device = dev->driver_data; | |
374 | struct net_device *netdev = device->dev; | |
375 | struct uml_net_private *lp = netdev->priv; | |
376 | ||
cd1ae0e4 | 377 | if (lp->remove != NULL) |
2e3f5251 JD |
378 | (*lp->remove)(&lp->user); |
379 | list_del(&device->list); | |
380 | kfree(device); | |
381 | free_netdev(netdev); | |
382 | } | |
383 | ||
80e39311 JD |
384 | /* |
385 | * Ensures that platform_driver_register is called only once by | |
386 | * eth_configure. Will be set in an initcall. | |
387 | */ | |
388 | static int driver_registered; | |
389 | ||
f34d9d2d JD |
390 | static void eth_configure(int n, void *init, char *mac, |
391 | struct transport *transport) | |
1da177e4 LT |
392 | { |
393 | struct uml_net *device; | |
394 | struct net_device *dev; | |
395 | struct uml_net_private *lp; | |
c74c69b4 | 396 | int err, size; |
1da177e4 | 397 | |
c74c69b4 | 398 | size = transport->private_size + sizeof(struct uml_net_private); |
1da177e4 | 399 | |
0268bd0a | 400 | device = kzalloc(sizeof(*device), GFP_KERNEL); |
1da177e4 | 401 | if (device == NULL) { |
8c840835 PBG |
402 | printk(KERN_ERR "eth_configure failed to allocate struct " |
403 | "uml_net\n"); | |
f34d9d2d | 404 | return; |
1da177e4 LT |
405 | } |
406 | ||
8c840835 PBG |
407 | dev = alloc_etherdev(size); |
408 | if (dev == NULL) { | |
409 | printk(KERN_ERR "eth_configure: failed to allocate struct " | |
410 | "net_device for eth%d\n", n); | |
411 | goto out_free_device; | |
412 | } | |
413 | ||
1da177e4 LT |
414 | INIT_LIST_HEAD(&device->list); |
415 | device->index = n; | |
416 | ||
e024715f PBG |
417 | /* If this name ends up conflicting with an existing registered |
418 | * netdevice, that is OK, register_netdev{,ice}() will notice this | |
419 | * and fail. | |
420 | */ | |
8c840835 | 421 | snprintf(dev->name, sizeof(dev->name), "eth%d", n); |
e024715f | 422 | |
8c840835 | 423 | setup_etheraddr(mac, device->mac, dev->name); |
1da177e4 LT |
424 | |
425 | printk(KERN_INFO "Netdevice %d ", n); | |
b10aeeef JD |
426 | printk("(%02x:%02x:%02x:%02x:%02x:%02x) ", |
427 | device->mac[0], device->mac[1], | |
428 | device->mac[2], device->mac[3], | |
429 | device->mac[4], device->mac[5]); | |
1da177e4 | 430 | printk(": "); |
1da177e4 | 431 | |
e56a7885 PBG |
432 | lp = dev->priv; |
433 | /* This points to the transport private data. It's still clear, but we | |
434 | * must memset it to 0 *now*. Let's help the drivers. */ | |
435 | memset(lp, 0, size); | |
eff3b634 | 436 | INIT_WORK(&lp->work, uml_dev_close); |
e56a7885 | 437 | |
1da177e4 LT |
438 | /* sysfs register */ |
439 | if (!driver_registered) { | |
3ae5eaec | 440 | platform_driver_register(¨_net_driver); |
1da177e4 LT |
441 | driver_registered = 1; |
442 | } | |
443 | device->pdev.id = n; | |
444 | device->pdev.name = DRIVER_NAME; | |
2e3f5251 JD |
445 | device->pdev.dev.release = net_device_release; |
446 | device->pdev.dev.driver_data = device; | |
cd1ae0e4 | 447 | if (platform_device_register(&device->pdev)) |
f34d9d2d | 448 | goto out_free_netdev; |
1da177e4 LT |
449 | SET_NETDEV_DEV(dev,&device->pdev.dev); |
450 | ||
1da177e4 LT |
451 | device->dev = dev; |
452 | ||
f34d9d2d JD |
453 | /* |
454 | * These just fill in a data structure, so there's no failure | |
455 | * to be worried about. | |
456 | */ | |
1da177e4 LT |
457 | (*transport->kern->init)(dev, init); |
458 | ||
1da177e4 LT |
459 | *lp = ((struct uml_net_private) |
460 | { .list = LIST_HEAD_INIT(lp->list), | |
461 | .dev = dev, | |
462 | .fd = -1, | |
463 | .mac = { 0xfe, 0xfd, 0x0, 0x0, 0x0, 0x0}, | |
b53f35a8 | 464 | .max_packet = transport->user->max_packet, |
1da177e4 LT |
465 | .protocol = transport->kern->protocol, |
466 | .open = transport->user->open, | |
467 | .close = transport->user->close, | |
468 | .remove = transport->user->remove, | |
469 | .read = transport->kern->read, | |
470 | .write = transport->kern->write, | |
471 | .add_address = transport->user->add_address, | |
b53f35a8 | 472 | .delete_address = transport->user->delete_address }); |
1da177e4 LT |
473 | |
474 | init_timer(&lp->tl); | |
475 | spin_lock_init(&lp->lock); | |
476 | lp->tl.function = uml_net_user_timer_expire; | |
b10aeeef | 477 | memcpy(lp->mac, device->mac, sizeof(lp->mac)); |
1da177e4 | 478 | |
f34d9d2d JD |
479 | if ((transport->user->init != NULL) && |
480 | ((*transport->user->init)(&lp->user, dev) != 0)) | |
481 | goto out_unregister; | |
1da177e4 | 482 | |
b10aeeef | 483 | set_ether_mac(dev, device->mac); |
b53f35a8 | 484 | dev->mtu = transport->user->mtu; |
f34d9d2d JD |
485 | dev->open = uml_net_open; |
486 | dev->hard_start_xmit = uml_net_start_xmit; | |
487 | dev->stop = uml_net_close; | |
488 | dev->get_stats = uml_net_get_stats; | |
489 | dev->set_multicast_list = uml_net_set_multicast_list; | |
490 | dev->tx_timeout = uml_net_tx_timeout; | |
491 | dev->set_mac_address = uml_net_set_mac; | |
492 | dev->change_mtu = uml_net_change_mtu; | |
493 | dev->ethtool_ops = ¨_net_ethtool_ops; | |
494 | dev->watchdog_timeo = (HZ >> 1); | |
495 | dev->irq = UM_ETH_IRQ; | |
1da177e4 | 496 | |
605c1e57 JD |
497 | err = update_drop_skb(lp->max_packet); |
498 | if (err) | |
499 | goto out_undo_user_init; | |
500 | ||
f34d9d2d JD |
501 | rtnl_lock(); |
502 | err = register_netdevice(dev); | |
503 | rtnl_unlock(); | |
504 | if (err) | |
505 | goto out_undo_user_init; | |
506 | ||
507 | spin_lock(&devices_lock); | |
508 | list_add(&device->list, &devices); | |
509 | spin_unlock(&devices_lock); | |
510 | ||
511 | return; | |
512 | ||
513 | out_undo_user_init: | |
8c840835 | 514 | if (transport->user->remove != NULL) |
f34d9d2d JD |
515 | (*transport->user->remove)(&lp->user); |
516 | out_unregister: | |
517 | platform_device_unregister(&device->pdev); | |
7d98230a | 518 | return; /* platform_device_unregister frees dev and device */ |
f34d9d2d JD |
519 | out_free_netdev: |
520 | free_netdev(dev); | |
8c840835 | 521 | out_free_device: |
f34d9d2d | 522 | kfree(device); |
1da177e4 LT |
523 | } |
524 | ||
525 | static struct uml_net *find_device(int n) | |
526 | { | |
527 | struct uml_net *device; | |
528 | struct list_head *ele; | |
529 | ||
530 | spin_lock(&devices_lock); | |
cd1ae0e4 | 531 | list_for_each(ele, &devices) { |
1da177e4 | 532 | device = list_entry(ele, struct uml_net, list); |
cd1ae0e4 | 533 | if (device->index == n) |
1da177e4 LT |
534 | goto out; |
535 | } | |
536 | device = NULL; | |
537 | out: | |
538 | spin_unlock(&devices_lock); | |
4ea21cd9 | 539 | return device; |
1da177e4 LT |
540 | } |
541 | ||
f28169d2 JD |
542 | static int eth_parse(char *str, int *index_out, char **str_out, |
543 | char **error_out) | |
1da177e4 LT |
544 | { |
545 | char *end; | |
f28169d2 | 546 | int n, err = -EINVAL;; |
1da177e4 LT |
547 | |
548 | n = simple_strtoul(str, &end, 0); | |
cd1ae0e4 | 549 | if (end == str) { |
f28169d2 JD |
550 | *error_out = "Bad device number"; |
551 | return err; | |
1da177e4 | 552 | } |
f28169d2 | 553 | |
1da177e4 | 554 | str = end; |
cd1ae0e4 | 555 | if (*str != '=') { |
f28169d2 JD |
556 | *error_out = "Expected '=' after device number"; |
557 | return err; | |
1da177e4 | 558 | } |
f28169d2 | 559 | |
1da177e4 | 560 | str++; |
cd1ae0e4 | 561 | if (find_device(n)) { |
f28169d2 JD |
562 | *error_out = "Device already configured"; |
563 | return err; | |
1da177e4 | 564 | } |
f28169d2 JD |
565 | |
566 | *index_out = n; | |
1da177e4 | 567 | *str_out = str; |
f28169d2 | 568 | return 0; |
1da177e4 LT |
569 | } |
570 | ||
571 | struct eth_init { | |
572 | struct list_head list; | |
573 | char *init; | |
574 | int index; | |
575 | }; | |
576 | ||
d3b7f69d JD |
577 | static DEFINE_SPINLOCK(transports_lock); |
578 | static LIST_HEAD(transports); | |
1da177e4 LT |
579 | |
580 | /* Filled in during early boot */ | |
c862fc32 | 581 | static LIST_HEAD(eth_cmd_line); |
1da177e4 LT |
582 | |
583 | static int check_transport(struct transport *transport, char *eth, int n, | |
584 | void **init_out, char **mac_out) | |
585 | { | |
586 | int len; | |
587 | ||
588 | len = strlen(transport->name); | |
cd1ae0e4 | 589 | if (strncmp(eth, transport->name, len)) |
4ea21cd9 | 590 | return 0; |
1da177e4 LT |
591 | |
592 | eth += len; | |
cd1ae0e4 | 593 | if (*eth == ',') |
1da177e4 | 594 | eth++; |
cd1ae0e4 | 595 | else if (*eth != '\0') |
4ea21cd9 | 596 | return 0; |
1da177e4 LT |
597 | |
598 | *init_out = kmalloc(transport->setup_size, GFP_KERNEL); | |
cd1ae0e4 | 599 | if (*init_out == NULL) |
4ea21cd9 | 600 | return 1; |
1da177e4 | 601 | |
cd1ae0e4 | 602 | if (!transport->setup(eth, mac_out, *init_out)) { |
1da177e4 LT |
603 | kfree(*init_out); |
604 | *init_out = NULL; | |
605 | } | |
4ea21cd9 | 606 | return 1; |
1da177e4 LT |
607 | } |
608 | ||
609 | void register_transport(struct transport *new) | |
610 | { | |
611 | struct list_head *ele, *next; | |
612 | struct eth_init *eth; | |
613 | void *init; | |
614 | char *mac = NULL; | |
615 | int match; | |
616 | ||
d3b7f69d JD |
617 | spin_lock(&transports_lock); |
618 | BUG_ON(!list_empty(&new->list)); | |
1da177e4 | 619 | list_add(&new->list, &transports); |
d3b7f69d | 620 | spin_unlock(&transports_lock); |
1da177e4 | 621 | |
cd1ae0e4 | 622 | list_for_each_safe(ele, next, ð_cmd_line) { |
1da177e4 LT |
623 | eth = list_entry(ele, struct eth_init, list); |
624 | match = check_transport(new, eth->init, eth->index, &init, | |
625 | &mac); | |
cd1ae0e4 | 626 | if (!match) |
1da177e4 | 627 | continue; |
cd1ae0e4 | 628 | else if (init != NULL) { |
1da177e4 LT |
629 | eth_configure(eth->index, init, mac, new); |
630 | kfree(init); | |
631 | } | |
632 | list_del(ð->list); | |
633 | } | |
634 | } | |
635 | ||
636 | static int eth_setup_common(char *str, int index) | |
637 | { | |
638 | struct list_head *ele; | |
639 | struct transport *transport; | |
640 | void *init; | |
641 | char *mac = NULL; | |
c862fc32 | 642 | int found = 0; |
1da177e4 | 643 | |
c862fc32 | 644 | spin_lock(&transports_lock); |
cd1ae0e4 | 645 | list_for_each(ele, &transports) { |
1da177e4 | 646 | transport = list_entry(ele, struct transport, list); |
cd1ae0e4 | 647 | if (!check_transport(transport, str, index, &init, &mac)) |
1da177e4 | 648 | continue; |
cd1ae0e4 | 649 | if (init != NULL) { |
1da177e4 LT |
650 | eth_configure(index, init, mac, transport); |
651 | kfree(init); | |
652 | } | |
c862fc32 JD |
653 | found = 1; |
654 | break; | |
1da177e4 | 655 | } |
c862fc32 JD |
656 | |
657 | spin_unlock(&transports_lock); | |
658 | return found; | |
1da177e4 LT |
659 | } |
660 | ||
97a1fcbb | 661 | static int __init eth_setup(char *str) |
1da177e4 LT |
662 | { |
663 | struct eth_init *new; | |
f28169d2 | 664 | char *error; |
1da177e4 LT |
665 | int n, err; |
666 | ||
f28169d2 | 667 | err = eth_parse(str, &n, &str, &error); |
cd1ae0e4 | 668 | if (err) { |
f28169d2 JD |
669 | printk(KERN_ERR "eth_setup - Couldn't parse '%s' : %s\n", |
670 | str, error); | |
1183dc94 | 671 | return 1; |
f28169d2 | 672 | } |
1da177e4 | 673 | |
1183dc94 | 674 | new = alloc_bootmem(sizeof(*new)); |
cd1ae0e4 JD |
675 | if (new == NULL) { |
676 | printk(KERN_ERR "eth_init : alloc_bootmem failed\n"); | |
1183dc94 | 677 | return 1; |
1da177e4 LT |
678 | } |
679 | ||
680 | INIT_LIST_HEAD(&new->list); | |
681 | new->index = n; | |
682 | new->init = str; | |
683 | ||
684 | list_add_tail(&new->list, ð_cmd_line); | |
1183dc94 | 685 | return 1; |
1da177e4 LT |
686 | } |
687 | ||
688 | __setup("eth", eth_setup); | |
689 | __uml_help(eth_setup, | |
690 | "eth[0-9]+=<transport>,<options>\n" | |
691 | " Configure a network device.\n\n" | |
692 | ); | |
693 | ||
f28169d2 | 694 | static int net_config(char *str, char **error_out) |
1da177e4 LT |
695 | { |
696 | int n, err; | |
697 | ||
f28169d2 | 698 | err = eth_parse(str, &n, &str, error_out); |
cd1ae0e4 | 699 | if (err) |
f28169d2 | 700 | return err; |
1da177e4 | 701 | |
f28169d2 JD |
702 | /* This string is broken up and the pieces used by the underlying |
703 | * driver. So, it is freed only if eth_setup_common fails. | |
704 | */ | |
970d6e3a | 705 | str = kstrdup(str, GFP_KERNEL); |
cd1ae0e4 | 706 | if (str == NULL) { |
f28169d2 JD |
707 | *error_out = "net_config failed to strdup string"; |
708 | return -ENOMEM; | |
1da177e4 LT |
709 | } |
710 | err = !eth_setup_common(str, n); | |
cd1ae0e4 | 711 | if (err) |
1da177e4 | 712 | kfree(str); |
cd1ae0e4 | 713 | return err; |
1da177e4 LT |
714 | } |
715 | ||
29d56cfe JD |
716 | static int net_id(char **str, int *start_out, int *end_out) |
717 | { | |
cd1ae0e4 JD |
718 | char *end; |
719 | int n; | |
29d56cfe JD |
720 | |
721 | n = simple_strtoul(*str, &end, 0); | |
cd1ae0e4 | 722 | if ((*end != '\0') || (end == *str)) |
29d56cfe JD |
723 | return -1; |
724 | ||
cd1ae0e4 JD |
725 | *start_out = n; |
726 | *end_out = n; | |
727 | *str = end; | |
728 | return n; | |
29d56cfe JD |
729 | } |
730 | ||
f28169d2 | 731 | static int net_remove(int n, char **error_out) |
1da177e4 LT |
732 | { |
733 | struct uml_net *device; | |
734 | struct net_device *dev; | |
735 | struct uml_net_private *lp; | |
1da177e4 LT |
736 | |
737 | device = find_device(n); | |
cd1ae0e4 | 738 | if (device == NULL) |
29d56cfe | 739 | return -ENODEV; |
1da177e4 LT |
740 | |
741 | dev = device->dev; | |
742 | lp = dev->priv; | |
cd1ae0e4 | 743 | if (lp->fd > 0) |
4ea21cd9 | 744 | return -EBUSY; |
1da177e4 LT |
745 | unregister_netdev(dev); |
746 | platform_device_unregister(&device->pdev); | |
747 | ||
29d56cfe | 748 | return 0; |
1da177e4 LT |
749 | } |
750 | ||
751 | static struct mc_device net_mc = { | |
84f48d4f | 752 | .list = LIST_HEAD_INIT(net_mc.list), |
1da177e4 LT |
753 | .name = "eth", |
754 | .config = net_config, | |
755 | .get_config = NULL, | |
4ea21cd9 | 756 | .id = net_id, |
1da177e4 LT |
757 | .remove = net_remove, |
758 | }; | |
759 | ||
32f862c3 | 760 | #ifdef CONFIG_INET |
1da177e4 LT |
761 | static int uml_inetaddr_event(struct notifier_block *this, unsigned long event, |
762 | void *ptr) | |
763 | { | |
764 | struct in_ifaddr *ifa = ptr; | |
1da177e4 LT |
765 | struct net_device *dev = ifa->ifa_dev->dev; |
766 | struct uml_net_private *lp; | |
767 | void (*proc)(unsigned char *, unsigned char *, void *); | |
768 | unsigned char addr_buf[4], netmask_buf[4]; | |
769 | ||
cd1ae0e4 | 770 | if (dev->open != uml_net_open) |
4ea21cd9 | 771 | return NOTIFY_DONE; |
1da177e4 LT |
772 | |
773 | lp = dev->priv; | |
774 | ||
775 | proc = NULL; | |
cd1ae0e4 | 776 | switch (event) { |
1da177e4 LT |
777 | case NETDEV_UP: |
778 | proc = lp->add_address; | |
779 | break; | |
780 | case NETDEV_DOWN: | |
781 | proc = lp->delete_address; | |
782 | break; | |
783 | } | |
cd1ae0e4 | 784 | if (proc != NULL) { |
0e76422c BS |
785 | memcpy(addr_buf, &ifa->ifa_address, sizeof(addr_buf)); |
786 | memcpy(netmask_buf, &ifa->ifa_mask, sizeof(netmask_buf)); | |
1da177e4 LT |
787 | (*proc)(addr_buf, netmask_buf, &lp->user); |
788 | } | |
4ea21cd9 | 789 | return NOTIFY_DONE; |
1da177e4 LT |
790 | } |
791 | ||
c862fc32 | 792 | /* uml_net_init shouldn't be called twice on two CPUs at the same time */ |
1da177e4 LT |
793 | struct notifier_block uml_inetaddr_notifier = { |
794 | .notifier_call = uml_inetaddr_event, | |
795 | }; | |
796 | ||
32f862c3 | 797 | static void inet_register(void) |
1da177e4 LT |
798 | { |
799 | struct list_head *ele; | |
f28169d2 | 800 | struct uml_net_private *lp; |
1da177e4 LT |
801 | struct in_device *ip; |
802 | struct in_ifaddr *in; | |
803 | ||
1da177e4 LT |
804 | register_inetaddr_notifier(¨_inetaddr_notifier); |
805 | ||
806 | /* Devices may have been opened already, so the uml_inetaddr_notifier | |
807 | * didn't get a chance to run for them. This fakes it so that | |
808 | * addresses which have already been set up get handled properly. | |
809 | */ | |
c862fc32 | 810 | spin_lock(&opened_lock); |
cd1ae0e4 | 811 | list_for_each(ele, &opened) { |
1da177e4 LT |
812 | lp = list_entry(ele, struct uml_net_private, list); |
813 | ip = lp->dev->ip_ptr; | |
cd1ae0e4 | 814 | if (ip == NULL) |
c862fc32 | 815 | continue; |
1da177e4 | 816 | in = ip->ifa_list; |
cd1ae0e4 | 817 | while (in != NULL) { |
1da177e4 LT |
818 | uml_inetaddr_event(NULL, NETDEV_UP, in); |
819 | in = in->ifa_next; | |
820 | } | |
f28169d2 | 821 | } |
c862fc32 | 822 | spin_unlock(&opened_lock); |
32f862c3 JD |
823 | } |
824 | #else | |
825 | static inline void inet_register(void) | |
826 | { | |
827 | } | |
828 | #endif | |
1da177e4 | 829 | |
32f862c3 JD |
830 | static int uml_net_init(void) |
831 | { | |
832 | mconsole_register_dev(&net_mc); | |
833 | inet_register(); | |
c862fc32 | 834 | return 0; |
1da177e4 LT |
835 | } |
836 | ||
837 | __initcall(uml_net_init); | |
838 | ||
839 | static void close_devices(void) | |
840 | { | |
841 | struct list_head *ele; | |
842 | struct uml_net_private *lp; | |
843 | ||
c862fc32 | 844 | spin_lock(&opened_lock); |
cd1ae0e4 | 845 | list_for_each(ele, &opened) { |
1da177e4 | 846 | lp = list_entry(ele, struct uml_net_private, list); |
8d93c700 | 847 | free_irq(lp->dev->irq, lp->dev); |
cd1ae0e4 | 848 | if ((lp->close != NULL) && (lp->fd >= 0)) |
1da177e4 | 849 | (*lp->close)(lp->fd, &lp->user); |
cd1ae0e4 | 850 | if (lp->remove != NULL) |
c862fc32 | 851 | (*lp->remove)(&lp->user); |
1da177e4 | 852 | } |
c862fc32 | 853 | spin_unlock(&opened_lock); |
1da177e4 LT |
854 | } |
855 | ||
856 | __uml_exitcall(close_devices); | |
857 | ||
f28169d2 JD |
858 | void iter_addresses(void *d, void (*cb)(unsigned char *, unsigned char *, |
859 | void *), | |
1da177e4 LT |
860 | void *arg) |
861 | { | |
862 | struct net_device *dev = d; | |
863 | struct in_device *ip = dev->ip_ptr; | |
864 | struct in_ifaddr *in; | |
865 | unsigned char address[4], netmask[4]; | |
866 | ||
cd1ae0e4 | 867 | if (ip == NULL) return; |
1da177e4 | 868 | in = ip->ifa_list; |
cd1ae0e4 | 869 | while (in != NULL) { |
0e76422c BS |
870 | memcpy(address, &in->ifa_address, sizeof(address)); |
871 | memcpy(netmask, &in->ifa_mask, sizeof(netmask)); | |
1da177e4 LT |
872 | (*cb)(address, netmask, arg); |
873 | in = in->ifa_next; | |
874 | } | |
875 | } | |
876 | ||
877 | int dev_netmask(void *d, void *m) | |
878 | { | |
879 | struct net_device *dev = d; | |
880 | struct in_device *ip = dev->ip_ptr; | |
881 | struct in_ifaddr *in; | |
a144ea4b | 882 | __be32 *mask_out = m; |
1da177e4 | 883 | |
cd1ae0e4 JD |
884 | if (ip == NULL) |
885 | return 1; | |
1da177e4 LT |
886 | |
887 | in = ip->ifa_list; | |
cd1ae0e4 JD |
888 | if (in == NULL) |
889 | return 1; | |
1da177e4 LT |
890 | |
891 | *mask_out = in->ifa_mask; | |
cd1ae0e4 | 892 | return 0; |
1da177e4 LT |
893 | } |
894 | ||
895 | void *get_output_buffer(int *len_out) | |
896 | { | |
897 | void *ret; | |
898 | ||
899 | ret = (void *) __get_free_pages(GFP_KERNEL, 0); | |
cd1ae0e4 | 900 | if (ret) *len_out = PAGE_SIZE; |
1da177e4 | 901 | else *len_out = 0; |
4ea21cd9 | 902 | return ret; |
1da177e4 LT |
903 | } |
904 | ||
905 | void free_output_buffer(void *buffer) | |
906 | { | |
907 | free_pages((unsigned long) buffer, 0); | |
908 | } | |
909 | ||
f28169d2 | 910 | int tap_setup_common(char *str, char *type, char **dev_name, char **mac_out, |
1da177e4 LT |
911 | char **gate_addr) |
912 | { | |
913 | char *remain; | |
914 | ||
915 | remain = split_if_spec(str, dev_name, mac_out, gate_addr, NULL); | |
cd1ae0e4 JD |
916 | if (remain != NULL) { |
917 | printk(KERN_ERR "tap_setup_common - Extra garbage on " | |
918 | "specification : '%s'\n", remain); | |
919 | return 1; | |
1da177e4 LT |
920 | } |
921 | ||
cd1ae0e4 | 922 | return 0; |
1da177e4 LT |
923 | } |
924 | ||
925 | unsigned short eth_protocol(struct sk_buff *skb) | |
926 | { | |
cd1ae0e4 | 927 | return eth_type_trans(skb, skb->dev); |
1da177e4 | 928 | } |