1 | // SPDX-License-Identifier: GPL-2.0-or-later |
2 | /* $Id: plip.c,v 1.3.6.2 1997/04/16 15:07:56 phil Exp $ */ |
3 | /* PLIP: A parallel port "network" driver for Linux. */ |
4 | /* This driver is for parallel port with 5-bit cable (LapLink (R) cable). */ |
5 | /* |
6 | * Authors: Donald Becker <becker@scyld.com> |
7 | * Tommy Thorn <thorn@daimi.aau.dk> |
8 | * Tanabe Hiroyasu <hiro@sanpo.t.u-tokyo.ac.jp> |
9 | * Alan Cox <gw4pts@gw4pts.ampr.org> |
10 | * Peter Bauer <100136.3530@compuserve.com> |
11 | * Niibe Yutaka <gniibe@mri.co.jp> |
12 | * Nimrod Zimerman <zimerman@mailandnews.com> |
13 | * |
14 | * Enhancements: |
15 | * Modularization and ifreq/ifmap support by Alan Cox. |
16 | * Rewritten by Niibe Yutaka. |
17 | * parport-sharing awareness code by Philip Blundell. |
18 | * SMP locking by Niibe Yutaka. |
19 | * Support for parallel ports with no IRQ (poll mode), |
20 | * Modifications to use the parallel port API |
21 | * by Nimrod Zimerman. |
22 | * |
23 | * Fixes: |
24 | * Niibe Yutaka |
25 | * - Module initialization. |
26 | * - MTU fix. |
27 | * - Make sure other end is OK, before sending a packet. |
28 | * - Fix immediate timer problem. |
29 | * |
30 | * Al Viro |
31 | * - Changed {enable,disable}_irq handling to make it work |
32 | * with new ("stack") semantics. |
33 | */ |
34 | |
35 | /* |
36 | * Original version and the name 'PLIP' from Donald Becker <becker@scyld.com> |
37 | * inspired by Russ Nelson's parallel port packet driver. |
38 | * |
39 | * NOTE: |
40 | * Tanabe Hiroyasu had changed the protocol, and it was in Linux v1.0. |
41 | * Because of the necessity to communicate to DOS machines with the |
42 | * Crynwr packet driver, Peter Bauer changed the protocol again |
43 | * back to original protocol. |
44 | * |
45 | * This version follows original PLIP protocol. |
46 | * So, this PLIP can't communicate the PLIP of Linux v1.0. |
47 | */ |
48 | |
49 | /* |
50 | * To use with DOS box, please do (Turn on ARP switch): |
51 | * # ifconfig plip[0-2] arp |
52 | */ |
53 | static const char version[] = "NET3 PLIP version 2.4-parport gniibe@mri.co.jp\n" ; |
54 | |
55 | /* |
56 | Sources: |
57 | Ideas and protocols came from Russ Nelson's <nelson@crynwr.com> |
58 | "parallel.asm" parallel port packet driver. |
59 | |
60 | The "Crynwr" parallel port standard specifies the following protocol: |
61 | Trigger by sending nibble '0x8' (this causes interrupt on other end) |
62 | count-low octet |
63 | count-high octet |
64 | ... data octets |
65 | checksum octet |
66 | Each octet is sent as <wait for rx. '0x1?'> <send 0x10+(octet&0x0F)> |
67 | <wait for rx. '0x0?'> <send 0x00+((octet>>4)&0x0F)> |
68 | |
69 | The packet is encapsulated as if it were ethernet. |
70 | |
71 | The cable used is a de facto standard parallel null cable -- sold as |
72 | a "LapLink" cable by various places. You'll need a 12-conductor cable to |
73 | make one yourself. The wiring is: |
74 | SLCTIN 17 - 17 |
75 | GROUND 25 - 25 |
76 | D0->ERROR 2 - 15 15 - 2 |
77 | D1->SLCT 3 - 13 13 - 3 |
78 | D2->PAPOUT 4 - 12 12 - 4 |
79 | D3->ACK 5 - 10 10 - 5 |
80 | D4->BUSY 6 - 11 11 - 6 |
81 | Do not connect the other pins. They are |
82 | D5,D6,D7 are 7,8,9 |
83 | STROBE is 1, FEED is 14, INIT is 16 |
84 | extra grounds are 18,19,20,21,22,23,24 |
85 | */ |
86 | |
87 | #include <linux/compat.h> |
88 | #include <linux/module.h> |
89 | #include <linux/kernel.h> |
90 | #include <linux/types.h> |
91 | #include <linux/fcntl.h> |
92 | #include <linux/interrupt.h> |
93 | #include <linux/string.h> |
94 | #include <linux/slab.h> |
95 | #include <linux/if_ether.h> |
96 | #include <linux/in.h> |
97 | #include <linux/errno.h> |
98 | #include <linux/delay.h> |
99 | #include <linux/init.h> |
100 | #include <linux/netdevice.h> |
101 | #include <linux/etherdevice.h> |
102 | #include <linux/inetdevice.h> |
103 | #include <linux/skbuff.h> |
104 | #include <linux/if_plip.h> |
105 | #include <linux/workqueue.h> |
106 | #include <linux/spinlock.h> |
107 | #include <linux/completion.h> |
108 | #include <linux/parport.h> |
109 | #include <linux/bitops.h> |
110 | |
111 | #include <net/neighbour.h> |
112 | |
113 | #include <asm/irq.h> |
114 | #include <asm/byteorder.h> |
115 | |
116 | /* Maximum number of devices to support. */ |
117 | #define PLIP_MAX 8 |
118 | |
119 | /* Use 0 for production, 1 for verification, >2 for debug */ |
120 | #ifndef NET_DEBUG |
121 | #define NET_DEBUG 1 |
122 | #endif |
123 | static const unsigned int net_debug = NET_DEBUG; |
124 | |
125 | #define ENABLE(irq) if (irq != -1) enable_irq(irq) |
126 | #define DISABLE(irq) if (irq != -1) disable_irq(irq) |
127 | |
128 | /* In micro second */ |
129 | #define PLIP_DELAY_UNIT 1 |
130 | |
131 | /* Connection time out = PLIP_TRIGGER_WAIT * PLIP_DELAY_UNIT usec */ |
132 | #define PLIP_TRIGGER_WAIT 500 |
133 | |
134 | /* Nibble time out = PLIP_NIBBLE_WAIT * PLIP_DELAY_UNIT usec */ |
135 | #define PLIP_NIBBLE_WAIT 3000 |
136 | |
137 | /* Bottom halves */ |
138 | static void plip_kick_bh(struct work_struct *work); |
139 | static void plip_bh(struct work_struct *work); |
140 | static void plip_timer_bh(struct work_struct *work); |
141 | |
142 | /* Interrupt handler */ |
143 | static void plip_interrupt(void *dev_id); |
144 | |
145 | /* Functions for DEV methods */ |
146 | static netdev_tx_t plip_tx_packet(struct sk_buff *skb, struct net_device *dev); |
147 | static int plip_hard_header(struct sk_buff *skb, struct net_device *dev, |
148 | unsigned short type, const void *daddr, |
149 | const void *saddr, unsigned len); |
150 | static int plip_hard_header_cache(const struct neighbour *neigh, |
151 | struct hh_cache *hh, __be16 type); |
152 | static int plip_open(struct net_device *dev); |
153 | static int plip_close(struct net_device *dev); |
154 | static int plip_siocdevprivate(struct net_device *dev, struct ifreq *ifr, |
155 | void __user *data, int cmd); |
156 | static int plip_preempt(void *handle); |
157 | static void plip_wakeup(void *handle); |
158 | |
159 | enum plip_connection_state { |
160 | PLIP_CN_NONE=0, |
161 | PLIP_CN_RECEIVE, |
162 | PLIP_CN_SEND, |
163 | PLIP_CN_CLOSING, |
164 | PLIP_CN_ERROR |
165 | }; |
166 | |
167 | enum plip_packet_state { |
168 | PLIP_PK_DONE=0, |
169 | PLIP_PK_TRIGGER, |
170 | PLIP_PK_LENGTH_LSB, |
171 | PLIP_PK_LENGTH_MSB, |
172 | PLIP_PK_DATA, |
173 | PLIP_PK_CHECKSUM |
174 | }; |
175 | |
176 | enum plip_nibble_state { |
177 | PLIP_NB_BEGIN, |
178 | PLIP_NB_1, |
179 | PLIP_NB_2, |
180 | }; |
181 | |
182 | struct plip_local { |
183 | enum plip_packet_state state; |
184 | enum plip_nibble_state nibble; |
185 | union { |
186 | struct { |
187 | #if defined(__LITTLE_ENDIAN) |
188 | unsigned char lsb; |
189 | unsigned char msb; |
190 | #elif defined(__BIG_ENDIAN) |
191 | unsigned char msb; |
192 | unsigned char lsb; |
193 | #else |
194 | #error "Please fix the endianness defines in <asm/byteorder.h>" |
195 | #endif |
196 | } b; |
197 | unsigned short h; |
198 | } length; |
199 | unsigned short byte; |
200 | unsigned char checksum; |
201 | unsigned char data; |
202 | struct sk_buff *skb; |
203 | }; |
204 | |
205 | struct net_local { |
206 | struct net_device *dev; |
207 | struct work_struct immediate; |
208 | struct delayed_work deferred; |
209 | struct delayed_work timer; |
210 | struct plip_local snd_data; |
211 | struct plip_local rcv_data; |
212 | struct pardevice *pardev; |
213 | unsigned long trigger; |
214 | unsigned long nibble; |
215 | enum plip_connection_state connection; |
216 | unsigned short timeout_count; |
217 | int is_deferred; |
218 | int port_owner; |
219 | int should_relinquish; |
220 | spinlock_t lock; |
221 | atomic_t kill_timer; |
222 | struct completion killed_timer_cmp; |
223 | }; |
224 | |
225 | static inline void enable_parport_interrupts (struct net_device *dev) |
226 | { |
227 | if (dev->irq != -1) |
228 | { |
229 | struct parport *port = |
230 | ((struct net_local *)netdev_priv(dev))->pardev->port; |
231 | port->ops->enable_irq (port); |
232 | } |
233 | } |
234 | |
235 | static inline void disable_parport_interrupts (struct net_device *dev) |
236 | { |
237 | if (dev->irq != -1) |
238 | { |
239 | struct parport *port = |
240 | ((struct net_local *)netdev_priv(dev))->pardev->port; |
241 | port->ops->disable_irq (port); |
242 | } |
243 | } |
244 | |
245 | static inline void write_data (struct net_device *dev, unsigned char data) |
246 | { |
247 | struct parport *port = |
248 | ((struct net_local *)netdev_priv(dev))->pardev->port; |
249 | |
250 | port->ops->write_data (port, data); |
251 | } |
252 | |
253 | static inline unsigned char read_status (struct net_device *dev) |
254 | { |
255 | struct parport *port = |
256 | ((struct net_local *)netdev_priv(dev))->pardev->port; |
257 | |
258 | return port->ops->read_status (port); |
259 | } |
260 | |
261 | static const struct header_ops = { |
262 | .create = plip_hard_header, |
263 | .cache = plip_hard_header_cache, |
264 | }; |
265 | |
266 | static const struct net_device_ops plip_netdev_ops = { |
267 | .ndo_open = plip_open, |
268 | .ndo_stop = plip_close, |
269 | .ndo_start_xmit = plip_tx_packet, |
270 | .ndo_siocdevprivate = plip_siocdevprivate, |
271 | .ndo_set_mac_address = eth_mac_addr, |
272 | .ndo_validate_addr = eth_validate_addr, |
273 | }; |
274 | |
275 | /* Entry point of PLIP driver. |
276 | Probe the hardware, and register/initialize the driver. |
277 | |
278 | PLIP is rather weird, because of the way it interacts with the parport |
279 | system. It is _not_ initialised from Space.c. Instead, plip_init() |
280 | is called, and that function makes up a "struct net_device" for each port, and |
281 | then calls us here. |
282 | |
283 | */ |
284 | static void |
285 | plip_init_netdev(struct net_device *dev) |
286 | { |
287 | static const u8 addr_init[ETH_ALEN] = { |
288 | 0xfc, 0xfc, 0xfc, |
289 | 0xfc, 0xfc, 0xfc, |
290 | }; |
291 | struct net_local *nl = netdev_priv(dev); |
292 | |
293 | /* Then, override parts of it */ |
294 | dev->tx_queue_len = 10; |
295 | dev->flags = IFF_POINTOPOINT|IFF_NOARP; |
296 | eth_hw_addr_set(dev, addr: addr_init); |
297 | |
298 | dev->netdev_ops = &plip_netdev_ops; |
299 | dev->header_ops = &plip_header_ops; |
300 | |
301 | |
302 | nl->port_owner = 0; |
303 | |
304 | /* Initialize constants */ |
305 | nl->trigger = PLIP_TRIGGER_WAIT; |
306 | nl->nibble = PLIP_NIBBLE_WAIT; |
307 | |
308 | /* Initialize task queue structures */ |
309 | INIT_WORK(&nl->immediate, plip_bh); |
310 | INIT_DELAYED_WORK(&nl->deferred, plip_kick_bh); |
311 | |
312 | if (dev->irq == -1) |
313 | INIT_DELAYED_WORK(&nl->timer, plip_timer_bh); |
314 | |
315 | spin_lock_init(&nl->lock); |
316 | } |
317 | |
318 | /* Bottom half handler for the delayed request. |
319 | This routine is kicked by do_timer(). |
320 | Request `plip_bh' to be invoked. */ |
321 | static void |
322 | plip_kick_bh(struct work_struct *work) |
323 | { |
324 | struct net_local *nl = |
325 | container_of(work, struct net_local, deferred.work); |
326 | |
327 | if (nl->is_deferred) |
328 | schedule_work(work: &nl->immediate); |
329 | } |
330 | |
331 | /* Forward declarations of internal routines */ |
332 | static int plip_none(struct net_device *, struct net_local *, |
333 | struct plip_local *, struct plip_local *); |
334 | static int plip_receive_packet(struct net_device *, struct net_local *, |
335 | struct plip_local *, struct plip_local *); |
336 | static int plip_send_packet(struct net_device *, struct net_local *, |
337 | struct plip_local *, struct plip_local *); |
338 | static int plip_connection_close(struct net_device *, struct net_local *, |
339 | struct plip_local *, struct plip_local *); |
340 | static int plip_error(struct net_device *, struct net_local *, |
341 | struct plip_local *, struct plip_local *); |
342 | static int plip_bh_timeout_error(struct net_device *dev, struct net_local *nl, |
343 | struct plip_local *snd, |
344 | struct plip_local *rcv, |
345 | int error); |
346 | |
347 | #define OK 0 |
348 | #define TIMEOUT 1 |
349 | #define ERROR 2 |
350 | #define HS_TIMEOUT 3 |
351 | |
352 | typedef int (*plip_func)(struct net_device *dev, struct net_local *nl, |
353 | struct plip_local *snd, struct plip_local *rcv); |
354 | |
355 | static const plip_func connection_state_table[] = |
356 | { |
357 | plip_none, |
358 | plip_receive_packet, |
359 | plip_send_packet, |
360 | plip_connection_close, |
361 | plip_error |
362 | }; |
363 | |
364 | /* Bottom half handler of PLIP. */ |
365 | static void |
366 | plip_bh(struct work_struct *work) |
367 | { |
368 | struct net_local *nl = container_of(work, struct net_local, immediate); |
369 | struct plip_local *snd = &nl->snd_data; |
370 | struct plip_local *rcv = &nl->rcv_data; |
371 | plip_func f; |
372 | int r; |
373 | |
374 | nl->is_deferred = 0; |
375 | f = connection_state_table[nl->connection]; |
376 | if ((r = (*f)(nl->dev, nl, snd, rcv)) != OK && |
377 | (r = plip_bh_timeout_error(dev: nl->dev, nl, snd, rcv, error: r)) != OK) { |
378 | nl->is_deferred = 1; |
379 | schedule_delayed_work(dwork: &nl->deferred, delay: 1); |
380 | } |
381 | } |
382 | |
383 | static void |
384 | plip_timer_bh(struct work_struct *work) |
385 | { |
386 | struct net_local *nl = |
387 | container_of(work, struct net_local, timer.work); |
388 | |
389 | if (!(atomic_read (v: &nl->kill_timer))) { |
390 | plip_interrupt (dev_id: nl->dev); |
391 | |
392 | schedule_delayed_work(dwork: &nl->timer, delay: 1); |
393 | } |
394 | else { |
395 | complete(&nl->killed_timer_cmp); |
396 | } |
397 | } |
398 | |
399 | static int |
400 | plip_bh_timeout_error(struct net_device *dev, struct net_local *nl, |
401 | struct plip_local *snd, struct plip_local *rcv, |
402 | int error) |
403 | { |
404 | unsigned char c0; |
405 | /* |
406 | * This is tricky. If we got here from the beginning of send (either |
407 | * with ERROR or HS_TIMEOUT) we have IRQ enabled. Otherwise it's |
408 | * already disabled. With the old variant of {enable,disable}_irq() |
409 | * extra disable_irq() was a no-op. Now it became mortal - it's |
410 | * unbalanced and thus we'll never re-enable IRQ (until rmmod plip, |
411 | * that is). So we have to treat HS_TIMEOUT and ERROR from send |
412 | * in a special way. |
413 | */ |
414 | |
415 | spin_lock_irq(lock: &nl->lock); |
416 | if (nl->connection == PLIP_CN_SEND) { |
417 | |
418 | if (error != ERROR) { /* Timeout */ |
419 | nl->timeout_count++; |
420 | if ((error == HS_TIMEOUT && nl->timeout_count <= 10) || |
421 | nl->timeout_count <= 3) { |
422 | spin_unlock_irq(lock: &nl->lock); |
423 | /* Try again later */ |
424 | return TIMEOUT; |
425 | } |
426 | c0 = read_status(dev); |
427 | printk(KERN_WARNING "%s: transmit timeout(%d,%02x)\n" , |
428 | dev->name, snd->state, c0); |
429 | } else |
430 | error = HS_TIMEOUT; |
431 | dev->stats.tx_errors++; |
432 | dev->stats.tx_aborted_errors++; |
433 | } else if (nl->connection == PLIP_CN_RECEIVE) { |
434 | if (rcv->state == PLIP_PK_TRIGGER) { |
435 | /* Transmission was interrupted. */ |
436 | spin_unlock_irq(lock: &nl->lock); |
437 | return OK; |
438 | } |
439 | if (error != ERROR) { /* Timeout */ |
440 | if (++nl->timeout_count <= 3) { |
441 | spin_unlock_irq(lock: &nl->lock); |
442 | /* Try again later */ |
443 | return TIMEOUT; |
444 | } |
445 | c0 = read_status(dev); |
446 | printk(KERN_WARNING "%s: receive timeout(%d,%02x)\n" , |
447 | dev->name, rcv->state, c0); |
448 | } |
449 | dev->stats.rx_dropped++; |
450 | } |
451 | rcv->state = PLIP_PK_DONE; |
452 | if (rcv->skb) { |
453 | dev_kfree_skb_irq(skb: rcv->skb); |
454 | rcv->skb = NULL; |
455 | } |
456 | snd->state = PLIP_PK_DONE; |
457 | if (snd->skb) { |
458 | dev_consume_skb_irq(skb: snd->skb); |
459 | snd->skb = NULL; |
460 | } |
461 | spin_unlock_irq(lock: &nl->lock); |
462 | if (error == HS_TIMEOUT) { |
463 | DISABLE(dev->irq); |
464 | synchronize_irq(irq: dev->irq); |
465 | } |
466 | disable_parport_interrupts (dev); |
467 | netif_stop_queue (dev); |
468 | nl->connection = PLIP_CN_ERROR; |
469 | write_data (dev, data: 0x00); |
470 | |
471 | return TIMEOUT; |
472 | } |
473 | |
474 | static int |
475 | plip_none(struct net_device *dev, struct net_local *nl, |
476 | struct plip_local *snd, struct plip_local *rcv) |
477 | { |
478 | return OK; |
479 | } |
480 | |
481 | /* PLIP_RECEIVE --- receive a byte(two nibbles) |
482 | Returns OK on success, TIMEOUT on timeout */ |
483 | static inline int |
484 | plip_receive(unsigned short nibble_timeout, struct net_device *dev, |
485 | enum plip_nibble_state *ns_p, unsigned char *data_p) |
486 | { |
487 | unsigned char c0, c1; |
488 | unsigned int cx; |
489 | |
490 | switch (*ns_p) { |
491 | case PLIP_NB_BEGIN: |
492 | cx = nibble_timeout; |
493 | while (1) { |
494 | c0 = read_status(dev); |
495 | udelay(PLIP_DELAY_UNIT); |
496 | if ((c0 & 0x80) == 0) { |
497 | c1 = read_status(dev); |
498 | if (c0 == c1) |
499 | break; |
500 | } |
501 | if (--cx == 0) |
502 | return TIMEOUT; |
503 | } |
504 | *data_p = (c0 >> 3) & 0x0f; |
505 | write_data (dev, data: 0x10); /* send ACK */ |
506 | *ns_p = PLIP_NB_1; |
507 | fallthrough; |
508 | |
509 | case PLIP_NB_1: |
510 | cx = nibble_timeout; |
511 | while (1) { |
512 | c0 = read_status(dev); |
513 | udelay(PLIP_DELAY_UNIT); |
514 | if (c0 & 0x80) { |
515 | c1 = read_status(dev); |
516 | if (c0 == c1) |
517 | break; |
518 | } |
519 | if (--cx == 0) |
520 | return TIMEOUT; |
521 | } |
522 | *data_p |= (c0 << 1) & 0xf0; |
523 | write_data (dev, data: 0x00); /* send ACK */ |
524 | *ns_p = PLIP_NB_BEGIN; |
525 | break; |
526 | case PLIP_NB_2: |
527 | break; |
528 | } |
529 | return OK; |
530 | } |
531 | |
532 | /* |
533 | * Determine the packet's protocol ID. The rule here is that we |
534 | * assume 802.3 if the type field is short enough to be a length. |
535 | * This is normal practice and works for any 'now in use' protocol. |
536 | * |
537 | * PLIP is ethernet ish but the daddr might not be valid if unicast. |
538 | * PLIP fortunately has no bus architecture (its Point-to-point). |
539 | * |
540 | * We can't fix the daddr thing as that quirk (more bug) is embedded |
541 | * in far too many old systems not all even running Linux. |
542 | */ |
543 | |
544 | static __be16 plip_type_trans(struct sk_buff *skb, struct net_device *dev) |
545 | { |
546 | struct ethhdr *eth; |
547 | unsigned char *rawp; |
548 | |
549 | skb_reset_mac_header(skb); |
550 | skb_pull(skb,len: dev->hard_header_len); |
551 | eth = eth_hdr(skb); |
552 | |
553 | if(is_multicast_ether_addr(addr: eth->h_dest)) |
554 | { |
555 | if(ether_addr_equal_64bits(addr1: eth->h_dest, addr2: dev->broadcast)) |
556 | skb->pkt_type=PACKET_BROADCAST; |
557 | else |
558 | skb->pkt_type=PACKET_MULTICAST; |
559 | } |
560 | |
561 | /* |
562 | * This ALLMULTI check should be redundant by 1.4 |
563 | * so don't forget to remove it. |
564 | */ |
565 | |
566 | if (ntohs(eth->h_proto) >= ETH_P_802_3_MIN) |
567 | return eth->h_proto; |
568 | |
569 | rawp = skb->data; |
570 | |
571 | /* |
572 | * This is a magic hack to spot IPX packets. Older Novell breaks |
573 | * the protocol design and runs IPX over 802.3 without an 802.2 LLC |
574 | * layer. We look for FFFF which isn't a used 802.2 SSAP/DSAP. This |
575 | * won't work for fault tolerant netware but does for the rest. |
576 | */ |
577 | if (*(unsigned short *)rawp == 0xFFFF) |
578 | return htons(ETH_P_802_3); |
579 | |
580 | /* |
581 | * Real 802.2 LLC |
582 | */ |
583 | return htons(ETH_P_802_2); |
584 | } |
585 | |
586 | /* PLIP_RECEIVE_PACKET --- receive a packet */ |
587 | static int |
588 | plip_receive_packet(struct net_device *dev, struct net_local *nl, |
589 | struct plip_local *snd, struct plip_local *rcv) |
590 | { |
591 | unsigned short nibble_timeout = nl->nibble; |
592 | unsigned char *lbuf; |
593 | |
594 | switch (rcv->state) { |
595 | case PLIP_PK_TRIGGER: |
596 | DISABLE(dev->irq); |
597 | /* Don't need to synchronize irq, as we can safely ignore it */ |
598 | disable_parport_interrupts (dev); |
599 | write_data (dev, data: 0x01); /* send ACK */ |
600 | if (net_debug > 2) |
601 | printk(KERN_DEBUG "%s: receive start\n" , dev->name); |
602 | rcv->state = PLIP_PK_LENGTH_LSB; |
603 | rcv->nibble = PLIP_NB_BEGIN; |
604 | fallthrough; |
605 | |
606 | case PLIP_PK_LENGTH_LSB: |
607 | if (snd->state != PLIP_PK_DONE) { |
608 | if (plip_receive(nibble_timeout: nl->trigger, dev, |
609 | ns_p: &rcv->nibble, data_p: &rcv->length.b.lsb)) { |
610 | /* collision, here dev->tbusy == 1 */ |
611 | rcv->state = PLIP_PK_DONE; |
612 | nl->is_deferred = 1; |
613 | nl->connection = PLIP_CN_SEND; |
614 | schedule_delayed_work(dwork: &nl->deferred, delay: 1); |
615 | enable_parport_interrupts (dev); |
616 | ENABLE(dev->irq); |
617 | return OK; |
618 | } |
619 | } else { |
620 | if (plip_receive(nibble_timeout, dev, |
621 | ns_p: &rcv->nibble, data_p: &rcv->length.b.lsb)) |
622 | return TIMEOUT; |
623 | } |
624 | rcv->state = PLIP_PK_LENGTH_MSB; |
625 | fallthrough; |
626 | |
627 | case PLIP_PK_LENGTH_MSB: |
628 | if (plip_receive(nibble_timeout, dev, |
629 | ns_p: &rcv->nibble, data_p: &rcv->length.b.msb)) |
630 | return TIMEOUT; |
631 | if (rcv->length.h > dev->mtu + dev->hard_header_len || |
632 | rcv->length.h < 8) { |
633 | printk(KERN_WARNING "%s: bogus packet size %d.\n" , dev->name, rcv->length.h); |
634 | return ERROR; |
635 | } |
636 | /* Malloc up new buffer. */ |
637 | rcv->skb = dev_alloc_skb(length: rcv->length.h + 2); |
638 | if (rcv->skb == NULL) { |
639 | printk(KERN_ERR "%s: Memory squeeze.\n" , dev->name); |
640 | return ERROR; |
641 | } |
642 | skb_reserve(skb: rcv->skb, len: 2); /* Align IP on 16 byte boundaries */ |
643 | skb_put(skb: rcv->skb,len: rcv->length.h); |
644 | rcv->skb->dev = dev; |
645 | rcv->state = PLIP_PK_DATA; |
646 | rcv->byte = 0; |
647 | rcv->checksum = 0; |
648 | fallthrough; |
649 | |
650 | case PLIP_PK_DATA: |
651 | lbuf = rcv->skb->data; |
652 | do { |
653 | if (plip_receive(nibble_timeout, dev, |
654 | ns_p: &rcv->nibble, data_p: &lbuf[rcv->byte])) |
655 | return TIMEOUT; |
656 | } while (++rcv->byte < rcv->length.h); |
657 | do { |
658 | rcv->checksum += lbuf[--rcv->byte]; |
659 | } while (rcv->byte); |
660 | rcv->state = PLIP_PK_CHECKSUM; |
661 | fallthrough; |
662 | |
663 | case PLIP_PK_CHECKSUM: |
664 | if (plip_receive(nibble_timeout, dev, |
665 | ns_p: &rcv->nibble, data_p: &rcv->data)) |
666 | return TIMEOUT; |
667 | if (rcv->data != rcv->checksum) { |
668 | dev->stats.rx_crc_errors++; |
669 | if (net_debug) |
670 | printk(KERN_DEBUG "%s: checksum error\n" , dev->name); |
671 | return ERROR; |
672 | } |
673 | rcv->state = PLIP_PK_DONE; |
674 | fallthrough; |
675 | |
676 | case PLIP_PK_DONE: |
677 | /* Inform the upper layer for the arrival of a packet. */ |
678 | rcv->skb->protocol=plip_type_trans(skb: rcv->skb, dev); |
679 | netif_rx(skb: rcv->skb); |
680 | dev->stats.rx_bytes += rcv->length.h; |
681 | dev->stats.rx_packets++; |
682 | rcv->skb = NULL; |
683 | if (net_debug > 2) |
684 | printk(KERN_DEBUG "%s: receive end\n" , dev->name); |
685 | |
686 | /* Close the connection. */ |
687 | write_data (dev, data: 0x00); |
688 | spin_lock_irq(lock: &nl->lock); |
689 | if (snd->state != PLIP_PK_DONE) { |
690 | nl->connection = PLIP_CN_SEND; |
691 | spin_unlock_irq(lock: &nl->lock); |
692 | schedule_work(work: &nl->immediate); |
693 | enable_parport_interrupts (dev); |
694 | ENABLE(dev->irq); |
695 | return OK; |
696 | } else { |
697 | nl->connection = PLIP_CN_NONE; |
698 | spin_unlock_irq(lock: &nl->lock); |
699 | enable_parport_interrupts (dev); |
700 | ENABLE(dev->irq); |
701 | return OK; |
702 | } |
703 | } |
704 | return OK; |
705 | } |
706 | |
707 | /* PLIP_SEND --- send a byte (two nibbles) |
708 | Returns OK on success, TIMEOUT when timeout */ |
709 | static inline int |
710 | plip_send(unsigned short nibble_timeout, struct net_device *dev, |
711 | enum plip_nibble_state *ns_p, unsigned char data) |
712 | { |
713 | unsigned char c0; |
714 | unsigned int cx; |
715 | |
716 | switch (*ns_p) { |
717 | case PLIP_NB_BEGIN: |
718 | write_data (dev, data: data & 0x0f); |
719 | *ns_p = PLIP_NB_1; |
720 | fallthrough; |
721 | |
722 | case PLIP_NB_1: |
723 | write_data (dev, data: 0x10 | (data & 0x0f)); |
724 | cx = nibble_timeout; |
725 | while (1) { |
726 | c0 = read_status(dev); |
727 | if ((c0 & 0x80) == 0) |
728 | break; |
729 | if (--cx == 0) |
730 | return TIMEOUT; |
731 | udelay(PLIP_DELAY_UNIT); |
732 | } |
733 | write_data (dev, data: 0x10 | (data >> 4)); |
734 | *ns_p = PLIP_NB_2; |
735 | fallthrough; |
736 | |
737 | case PLIP_NB_2: |
738 | write_data (dev, data: (data >> 4)); |
739 | cx = nibble_timeout; |
740 | while (1) { |
741 | c0 = read_status(dev); |
742 | if (c0 & 0x80) |
743 | break; |
744 | if (--cx == 0) |
745 | return TIMEOUT; |
746 | udelay(PLIP_DELAY_UNIT); |
747 | } |
748 | *ns_p = PLIP_NB_BEGIN; |
749 | return OK; |
750 | } |
751 | return OK; |
752 | } |
753 | |
754 | /* PLIP_SEND_PACKET --- send a packet */ |
755 | static int |
756 | plip_send_packet(struct net_device *dev, struct net_local *nl, |
757 | struct plip_local *snd, struct plip_local *rcv) |
758 | { |
759 | unsigned short nibble_timeout = nl->nibble; |
760 | unsigned char *lbuf; |
761 | unsigned char c0; |
762 | unsigned int cx; |
763 | |
764 | if (snd->skb == NULL || (lbuf = snd->skb->data) == NULL) { |
765 | printk(KERN_DEBUG "%s: send skb lost\n" , dev->name); |
766 | snd->state = PLIP_PK_DONE; |
767 | snd->skb = NULL; |
768 | return ERROR; |
769 | } |
770 | |
771 | switch (snd->state) { |
772 | case PLIP_PK_TRIGGER: |
773 | if ((read_status(dev) & 0xf8) != 0x80) |
774 | return HS_TIMEOUT; |
775 | |
776 | /* Trigger remote rx interrupt. */ |
777 | write_data (dev, data: 0x08); |
778 | cx = nl->trigger; |
779 | while (1) { |
780 | udelay(PLIP_DELAY_UNIT); |
781 | spin_lock_irq(lock: &nl->lock); |
782 | if (nl->connection == PLIP_CN_RECEIVE) { |
783 | spin_unlock_irq(lock: &nl->lock); |
784 | /* Interrupted. */ |
785 | dev->stats.collisions++; |
786 | return OK; |
787 | } |
788 | c0 = read_status(dev); |
789 | if (c0 & 0x08) { |
790 | spin_unlock_irq(lock: &nl->lock); |
791 | DISABLE(dev->irq); |
792 | synchronize_irq(irq: dev->irq); |
793 | if (nl->connection == PLIP_CN_RECEIVE) { |
794 | /* Interrupted. |
795 | We don't need to enable irq, |
796 | as it is soon disabled. */ |
797 | /* Yes, we do. New variant of |
798 | {enable,disable}_irq *counts* |
799 | them. -- AV */ |
800 | ENABLE(dev->irq); |
801 | dev->stats.collisions++; |
802 | return OK; |
803 | } |
804 | disable_parport_interrupts (dev); |
805 | if (net_debug > 2) |
806 | printk(KERN_DEBUG "%s: send start\n" , dev->name); |
807 | snd->state = PLIP_PK_LENGTH_LSB; |
808 | snd->nibble = PLIP_NB_BEGIN; |
809 | nl->timeout_count = 0; |
810 | break; |
811 | } |
812 | spin_unlock_irq(lock: &nl->lock); |
813 | if (--cx == 0) { |
814 | write_data (dev, data: 0x00); |
815 | return HS_TIMEOUT; |
816 | } |
817 | } |
818 | break; |
819 | |
820 | case PLIP_PK_LENGTH_LSB: |
821 | if (plip_send(nibble_timeout, dev, |
822 | ns_p: &snd->nibble, data: snd->length.b.lsb)) |
823 | return TIMEOUT; |
824 | snd->state = PLIP_PK_LENGTH_MSB; |
825 | fallthrough; |
826 | |
827 | case PLIP_PK_LENGTH_MSB: |
828 | if (plip_send(nibble_timeout, dev, |
829 | ns_p: &snd->nibble, data: snd->length.b.msb)) |
830 | return TIMEOUT; |
831 | snd->state = PLIP_PK_DATA; |
832 | snd->byte = 0; |
833 | snd->checksum = 0; |
834 | fallthrough; |
835 | |
836 | case PLIP_PK_DATA: |
837 | do { |
838 | if (plip_send(nibble_timeout, dev, |
839 | ns_p: &snd->nibble, data: lbuf[snd->byte])) |
840 | return TIMEOUT; |
841 | } while (++snd->byte < snd->length.h); |
842 | do { |
843 | snd->checksum += lbuf[--snd->byte]; |
844 | } while (snd->byte); |
845 | snd->state = PLIP_PK_CHECKSUM; |
846 | fallthrough; |
847 | |
848 | case PLIP_PK_CHECKSUM: |
849 | if (plip_send(nibble_timeout, dev, |
850 | ns_p: &snd->nibble, data: snd->checksum)) |
851 | return TIMEOUT; |
852 | |
853 | dev->stats.tx_bytes += snd->skb->len; |
854 | dev_kfree_skb(snd->skb); |
855 | dev->stats.tx_packets++; |
856 | snd->state = PLIP_PK_DONE; |
857 | fallthrough; |
858 | |
859 | case PLIP_PK_DONE: |
860 | /* Close the connection */ |
861 | write_data (dev, data: 0x00); |
862 | snd->skb = NULL; |
863 | if (net_debug > 2) |
864 | printk(KERN_DEBUG "%s: send end\n" , dev->name); |
865 | nl->connection = PLIP_CN_CLOSING; |
866 | nl->is_deferred = 1; |
867 | schedule_delayed_work(dwork: &nl->deferred, delay: 1); |
868 | enable_parport_interrupts (dev); |
869 | ENABLE(dev->irq); |
870 | return OK; |
871 | } |
872 | return OK; |
873 | } |
874 | |
875 | static int |
876 | plip_connection_close(struct net_device *dev, struct net_local *nl, |
877 | struct plip_local *snd, struct plip_local *rcv) |
878 | { |
879 | spin_lock_irq(lock: &nl->lock); |
880 | if (nl->connection == PLIP_CN_CLOSING) { |
881 | nl->connection = PLIP_CN_NONE; |
882 | netif_wake_queue (dev); |
883 | } |
884 | spin_unlock_irq(lock: &nl->lock); |
885 | if (nl->should_relinquish) { |
886 | nl->should_relinquish = nl->port_owner = 0; |
887 | parport_release(dev: nl->pardev); |
888 | } |
889 | return OK; |
890 | } |
891 | |
892 | /* PLIP_ERROR --- wait till other end settled */ |
893 | static int |
894 | plip_error(struct net_device *dev, struct net_local *nl, |
895 | struct plip_local *snd, struct plip_local *rcv) |
896 | { |
897 | unsigned char status; |
898 | |
899 | status = read_status(dev); |
900 | if ((status & 0xf8) == 0x80) { |
901 | if (net_debug > 2) |
902 | printk(KERN_DEBUG "%s: reset interface.\n" , dev->name); |
903 | nl->connection = PLIP_CN_NONE; |
904 | nl->should_relinquish = 0; |
905 | netif_start_queue (dev); |
906 | enable_parport_interrupts (dev); |
907 | ENABLE(dev->irq); |
908 | netif_wake_queue (dev); |
909 | } else { |
910 | nl->is_deferred = 1; |
911 | schedule_delayed_work(dwork: &nl->deferred, delay: 1); |
912 | } |
913 | |
914 | return OK; |
915 | } |
916 | |
917 | /* Handle the parallel port interrupts. */ |
918 | static void |
919 | plip_interrupt(void *dev_id) |
920 | { |
921 | struct net_device *dev = dev_id; |
922 | struct net_local *nl; |
923 | struct plip_local *rcv; |
924 | unsigned char c0; |
925 | unsigned long flags; |
926 | |
927 | nl = netdev_priv(dev); |
928 | rcv = &nl->rcv_data; |
929 | |
930 | spin_lock_irqsave (&nl->lock, flags); |
931 | |
932 | c0 = read_status(dev); |
933 | if ((c0 & 0xf8) != 0xc0) { |
934 | if ((dev->irq != -1) && (net_debug > 1)) |
935 | printk(KERN_DEBUG "%s: spurious interrupt\n" , dev->name); |
936 | spin_unlock_irqrestore (lock: &nl->lock, flags); |
937 | return; |
938 | } |
939 | |
940 | if (net_debug > 3) |
941 | printk(KERN_DEBUG "%s: interrupt.\n" , dev->name); |
942 | |
943 | switch (nl->connection) { |
944 | case PLIP_CN_CLOSING: |
945 | netif_wake_queue (dev); |
946 | fallthrough; |
947 | case PLIP_CN_NONE: |
948 | case PLIP_CN_SEND: |
949 | rcv->state = PLIP_PK_TRIGGER; |
950 | nl->connection = PLIP_CN_RECEIVE; |
951 | nl->timeout_count = 0; |
952 | schedule_work(work: &nl->immediate); |
953 | break; |
954 | |
955 | case PLIP_CN_RECEIVE: |
956 | /* May occur because there is race condition |
957 | around test and set of dev->interrupt. |
958 | Ignore this interrupt. */ |
959 | break; |
960 | |
961 | case PLIP_CN_ERROR: |
962 | printk(KERN_ERR "%s: receive interrupt in error state\n" , dev->name); |
963 | break; |
964 | } |
965 | |
966 | spin_unlock_irqrestore(lock: &nl->lock, flags); |
967 | } |
968 | |
969 | static netdev_tx_t |
970 | plip_tx_packet(struct sk_buff *skb, struct net_device *dev) |
971 | { |
972 | struct net_local *nl = netdev_priv(dev); |
973 | struct plip_local *snd = &nl->snd_data; |
974 | |
975 | if (netif_queue_stopped(dev)) |
976 | return NETDEV_TX_BUSY; |
977 | |
978 | /* We may need to grab the bus */ |
979 | if (!nl->port_owner) { |
980 | if (parport_claim(dev: nl->pardev)) |
981 | return NETDEV_TX_BUSY; |
982 | nl->port_owner = 1; |
983 | } |
984 | |
985 | netif_stop_queue (dev); |
986 | |
987 | if (skb->len > dev->mtu + dev->hard_header_len) { |
988 | printk(KERN_WARNING "%s: packet too big, %d.\n" , dev->name, (int)skb->len); |
989 | netif_start_queue (dev); |
990 | return NETDEV_TX_BUSY; |
991 | } |
992 | |
993 | if (net_debug > 2) |
994 | printk(KERN_DEBUG "%s: send request\n" , dev->name); |
995 | |
996 | spin_lock_irq(lock: &nl->lock); |
997 | snd->skb = skb; |
998 | snd->length.h = skb->len; |
999 | snd->state = PLIP_PK_TRIGGER; |
1000 | if (nl->connection == PLIP_CN_NONE) { |
1001 | nl->connection = PLIP_CN_SEND; |
1002 | nl->timeout_count = 0; |
1003 | } |
1004 | schedule_work(work: &nl->immediate); |
1005 | spin_unlock_irq(lock: &nl->lock); |
1006 | |
1007 | return NETDEV_TX_OK; |
1008 | } |
1009 | |
1010 | static void |
1011 | plip_rewrite_address(const struct net_device *dev, struct ethhdr *eth) |
1012 | { |
1013 | const struct in_device *in_dev; |
1014 | |
1015 | rcu_read_lock(); |
1016 | in_dev = __in_dev_get_rcu(dev); |
1017 | if (in_dev) { |
1018 | /* Any address will do - we take the first */ |
1019 | const struct in_ifaddr *ifa = rcu_dereference(in_dev->ifa_list); |
1020 | if (ifa) { |
1021 | memcpy(eth->h_source, dev->dev_addr, ETH_ALEN); |
1022 | memset(eth->h_dest, 0xfc, 2); |
1023 | memcpy(eth->h_dest+2, &ifa->ifa_address, 4); |
1024 | } |
1025 | } |
1026 | rcu_read_unlock(); |
1027 | } |
1028 | |
1029 | static int |
1030 | (struct sk_buff *skb, struct net_device *dev, |
1031 | unsigned short type, const void *daddr, |
1032 | const void *saddr, unsigned len) |
1033 | { |
1034 | int ret; |
1035 | |
1036 | ret = eth_header(skb, dev, type, daddr, saddr, len); |
1037 | if (ret >= 0) |
1038 | plip_rewrite_address (dev, eth: (struct ethhdr *)skb->data); |
1039 | |
1040 | return ret; |
1041 | } |
1042 | |
1043 | static int (const struct neighbour *neigh, |
1044 | struct hh_cache *hh, __be16 type) |
1045 | { |
1046 | int ret; |
1047 | |
1048 | ret = eth_header_cache(neigh, hh, type); |
1049 | if (ret == 0) { |
1050 | struct ethhdr *eth; |
1051 | |
1052 | eth = (struct ethhdr*)(((u8*)hh->hh_data) + |
1053 | HH_DATA_OFF(sizeof(*eth))); |
1054 | plip_rewrite_address (dev: neigh->dev, eth); |
1055 | } |
1056 | |
1057 | return ret; |
1058 | } |
1059 | |
1060 | /* Open/initialize the board. This is called (in the current kernel) |
1061 | sometime after booting when the 'ifconfig' program is run. |
1062 | |
1063 | This routine gets exclusive access to the parallel port by allocating |
1064 | its IRQ line. |
1065 | */ |
1066 | static int |
1067 | plip_open(struct net_device *dev) |
1068 | { |
1069 | struct net_local *nl = netdev_priv(dev); |
1070 | struct in_device *in_dev; |
1071 | |
1072 | /* Grab the port */ |
1073 | if (!nl->port_owner) { |
1074 | if (parport_claim(dev: nl->pardev)) return -EAGAIN; |
1075 | nl->port_owner = 1; |
1076 | } |
1077 | |
1078 | nl->should_relinquish = 0; |
1079 | |
1080 | /* Clear the data port. */ |
1081 | write_data (dev, data: 0x00); |
1082 | |
1083 | /* Enable rx interrupt. */ |
1084 | enable_parport_interrupts (dev); |
1085 | if (dev->irq == -1) |
1086 | { |
1087 | atomic_set (v: &nl->kill_timer, i: 0); |
1088 | schedule_delayed_work(dwork: &nl->timer, delay: 1); |
1089 | } |
1090 | |
1091 | /* Initialize the state machine. */ |
1092 | nl->rcv_data.state = nl->snd_data.state = PLIP_PK_DONE; |
1093 | nl->rcv_data.skb = nl->snd_data.skb = NULL; |
1094 | nl->connection = PLIP_CN_NONE; |
1095 | nl->is_deferred = 0; |
1096 | |
1097 | /* Fill in the MAC-level header. |
1098 | We used to abuse dev->broadcast to store the point-to-point |
1099 | MAC address, but we no longer do it. Instead, we fetch the |
1100 | interface address whenever it is needed, which is cheap enough |
1101 | because we use the hh_cache. Actually, abusing dev->broadcast |
1102 | didn't work, because when using plip_open the point-to-point |
1103 | address isn't yet known. |
1104 | PLIP doesn't have a real MAC address, but we need it to be |
1105 | DOS compatible, and to properly support taps (otherwise, |
1106 | when the device address isn't identical to the address of a |
1107 | received frame, the kernel incorrectly drops it). */ |
1108 | |
1109 | in_dev=__in_dev_get_rtnl(dev); |
1110 | if (in_dev) { |
1111 | /* Any address will do - we take the first. We already |
1112 | have the first two bytes filled with 0xfc, from |
1113 | plip_init_dev(). */ |
1114 | const struct in_ifaddr *ifa = rtnl_dereference(in_dev->ifa_list); |
1115 | if (ifa != NULL) { |
1116 | dev_addr_mod(dev, offset: 2, addr: &ifa->ifa_local, len: 4); |
1117 | } |
1118 | } |
1119 | |
1120 | netif_start_queue (dev); |
1121 | |
1122 | return 0; |
1123 | } |
1124 | |
1125 | /* The inverse routine to plip_open (). */ |
1126 | static int |
1127 | plip_close(struct net_device *dev) |
1128 | { |
1129 | struct net_local *nl = netdev_priv(dev); |
1130 | struct plip_local *snd = &nl->snd_data; |
1131 | struct plip_local *rcv = &nl->rcv_data; |
1132 | |
1133 | netif_stop_queue (dev); |
1134 | DISABLE(dev->irq); |
1135 | synchronize_irq(irq: dev->irq); |
1136 | |
1137 | if (dev->irq == -1) |
1138 | { |
1139 | init_completion(x: &nl->killed_timer_cmp); |
1140 | atomic_set (v: &nl->kill_timer, i: 1); |
1141 | wait_for_completion(&nl->killed_timer_cmp); |
1142 | } |
1143 | |
1144 | #ifdef NOTDEF |
1145 | outb(0x00, PAR_DATA(dev)); |
1146 | #endif |
1147 | nl->is_deferred = 0; |
1148 | nl->connection = PLIP_CN_NONE; |
1149 | if (nl->port_owner) { |
1150 | parport_release(dev: nl->pardev); |
1151 | nl->port_owner = 0; |
1152 | } |
1153 | |
1154 | snd->state = PLIP_PK_DONE; |
1155 | if (snd->skb) { |
1156 | dev_kfree_skb(snd->skb); |
1157 | snd->skb = NULL; |
1158 | } |
1159 | rcv->state = PLIP_PK_DONE; |
1160 | if (rcv->skb) { |
1161 | kfree_skb(skb: rcv->skb); |
1162 | rcv->skb = NULL; |
1163 | } |
1164 | |
1165 | #ifdef NOTDEF |
1166 | /* Reset. */ |
1167 | outb(0x00, PAR_CONTROL(dev)); |
1168 | #endif |
1169 | return 0; |
1170 | } |
1171 | |
1172 | static int |
1173 | plip_preempt(void *handle) |
1174 | { |
1175 | struct net_device *dev = (struct net_device *)handle; |
1176 | struct net_local *nl = netdev_priv(dev); |
1177 | |
1178 | /* Stand our ground if a datagram is on the wire */ |
1179 | if (nl->connection != PLIP_CN_NONE) { |
1180 | nl->should_relinquish = 1; |
1181 | return 1; |
1182 | } |
1183 | |
1184 | nl->port_owner = 0; /* Remember that we released the bus */ |
1185 | return 0; |
1186 | } |
1187 | |
1188 | static void |
1189 | plip_wakeup(void *handle) |
1190 | { |
1191 | struct net_device *dev = (struct net_device *)handle; |
1192 | struct net_local *nl = netdev_priv(dev); |
1193 | |
1194 | if (nl->port_owner) { |
1195 | /* Why are we being woken up? */ |
1196 | printk(KERN_DEBUG "%s: why am I being woken up?\n" , dev->name); |
1197 | if (!parport_claim(dev: nl->pardev)) |
1198 | /* bus_owner is already set (but why?) */ |
1199 | printk(KERN_DEBUG "%s: I'm broken.\n" , dev->name); |
1200 | else |
1201 | return; |
1202 | } |
1203 | |
1204 | if (!(dev->flags & IFF_UP)) |
1205 | /* Don't need the port when the interface is down */ |
1206 | return; |
1207 | |
1208 | if (!parport_claim(dev: nl->pardev)) { |
1209 | nl->port_owner = 1; |
1210 | /* Clear the data port. */ |
1211 | write_data (dev, data: 0x00); |
1212 | } |
1213 | } |
1214 | |
1215 | static int |
1216 | plip_siocdevprivate(struct net_device *dev, struct ifreq *rq, |
1217 | void __user *data, int cmd) |
1218 | { |
1219 | struct net_local *nl = netdev_priv(dev); |
1220 | struct plipconf *pc = (struct plipconf *) &rq->ifr_ifru; |
1221 | |
1222 | if (cmd != SIOCDEVPLIP) |
1223 | return -EOPNOTSUPP; |
1224 | |
1225 | if (in_compat_syscall()) |
1226 | return -EOPNOTSUPP; |
1227 | |
1228 | switch(pc->pcmd) { |
1229 | case PLIP_GET_TIMEOUT: |
1230 | pc->trigger = nl->trigger; |
1231 | pc->nibble = nl->nibble; |
1232 | break; |
1233 | case PLIP_SET_TIMEOUT: |
1234 | if(!capable(CAP_NET_ADMIN)) |
1235 | return -EPERM; |
1236 | nl->trigger = pc->trigger; |
1237 | nl->nibble = pc->nibble; |
1238 | break; |
1239 | default: |
1240 | return -EOPNOTSUPP; |
1241 | } |
1242 | return 0; |
1243 | } |
1244 | |
1245 | static int parport[PLIP_MAX] = { [0 ... PLIP_MAX-1] = -1 }; |
1246 | static int timid; |
1247 | |
1248 | module_param_array(parport, int, NULL, 0); |
1249 | module_param(timid, int, 0); |
1250 | MODULE_PARM_DESC(parport, "List of parport device numbers to use by plip" ); |
1251 | |
1252 | static struct net_device *dev_plip[PLIP_MAX] = { NULL, }; |
1253 | |
1254 | static inline int |
1255 | plip_searchfor(int list[], int a) |
1256 | { |
1257 | int i; |
1258 | for (i = 0; i < PLIP_MAX && list[i] != -1; i++) { |
1259 | if (list[i] == a) return 1; |
1260 | } |
1261 | return 0; |
1262 | } |
1263 | |
1264 | /* plip_attach() is called (by the parport code) when a port is |
1265 | * available to use. */ |
1266 | static void plip_attach (struct parport *port) |
1267 | { |
1268 | static int unit; |
1269 | struct net_device *dev; |
1270 | struct net_local *nl; |
1271 | char name[IFNAMSIZ]; |
1272 | struct pardev_cb plip_cb; |
1273 | |
1274 | if ((parport[0] == -1 && (!timid || !port->devices)) || |
1275 | plip_searchfor(list: parport, a: port->number)) { |
1276 | if (unit == PLIP_MAX) { |
1277 | printk(KERN_ERR "plip: too many devices\n" ); |
1278 | return; |
1279 | } |
1280 | |
1281 | sprintf(buf: name, fmt: "plip%d" , unit); |
1282 | dev = alloc_etherdev(sizeof(struct net_local)); |
1283 | if (!dev) |
1284 | return; |
1285 | |
1286 | strcpy(p: dev->name, q: name); |
1287 | |
1288 | dev->irq = port->irq; |
1289 | dev->base_addr = port->base; |
1290 | if (port->irq == -1) { |
1291 | printk(KERN_INFO "plip: %s has no IRQ. Using IRQ-less mode," |
1292 | "which is fairly inefficient!\n" , port->name); |
1293 | } |
1294 | |
1295 | nl = netdev_priv(dev); |
1296 | nl->dev = dev; |
1297 | |
1298 | memset(&plip_cb, 0, sizeof(plip_cb)); |
1299 | plip_cb.private = dev; |
1300 | plip_cb.preempt = plip_preempt; |
1301 | plip_cb.wakeup = plip_wakeup; |
1302 | plip_cb.irq_func = plip_interrupt; |
1303 | |
1304 | nl->pardev = parport_register_dev_model(port, name: dev->name, |
1305 | par_dev_cb: &plip_cb, cnt: unit); |
1306 | |
1307 | if (!nl->pardev) { |
1308 | printk(KERN_ERR "%s: parport_register failed\n" , name); |
1309 | goto err_free_dev; |
1310 | } |
1311 | |
1312 | plip_init_netdev(dev); |
1313 | |
1314 | if (register_netdev(dev)) { |
1315 | printk(KERN_ERR "%s: network register failed\n" , name); |
1316 | goto err_parport_unregister; |
1317 | } |
1318 | |
1319 | printk(KERN_INFO "%s" , version); |
1320 | if (dev->irq != -1) |
1321 | printk(KERN_INFO "%s: Parallel port at %#3lx, " |
1322 | "using IRQ %d.\n" , |
1323 | dev->name, dev->base_addr, dev->irq); |
1324 | else |
1325 | printk(KERN_INFO "%s: Parallel port at %#3lx, " |
1326 | "not using IRQ.\n" , |
1327 | dev->name, dev->base_addr); |
1328 | dev_plip[unit++] = dev; |
1329 | } |
1330 | return; |
1331 | |
1332 | err_parport_unregister: |
1333 | parport_unregister_device(dev: nl->pardev); |
1334 | err_free_dev: |
1335 | free_netdev(dev); |
1336 | } |
1337 | |
1338 | /* plip_detach() is called (by the parport code) when a port is |
1339 | * no longer available to use. */ |
1340 | static void plip_detach (struct parport *port) |
1341 | { |
1342 | /* Nothing to do */ |
1343 | } |
1344 | |
1345 | static int plip_probe(struct pardevice *par_dev) |
1346 | { |
1347 | struct device_driver *drv = par_dev->dev.driver; |
1348 | int len = strlen(drv->name); |
1349 | |
1350 | if (strncmp(par_dev->name, drv->name, len)) |
1351 | return -ENODEV; |
1352 | |
1353 | return 0; |
1354 | } |
1355 | |
1356 | static struct parport_driver plip_driver = { |
1357 | .name = "plip" , |
1358 | .probe = plip_probe, |
1359 | .match_port = plip_attach, |
1360 | .detach = plip_detach, |
1361 | .devmodel = true, |
1362 | }; |
1363 | |
1364 | static void __exit plip_cleanup_module (void) |
1365 | { |
1366 | struct net_device *dev; |
1367 | int i; |
1368 | |
1369 | for (i=0; i < PLIP_MAX; i++) { |
1370 | if ((dev = dev_plip[i])) { |
1371 | struct net_local *nl = netdev_priv(dev); |
1372 | unregister_netdev(dev); |
1373 | if (nl->port_owner) |
1374 | parport_release(dev: nl->pardev); |
1375 | parport_unregister_device(dev: nl->pardev); |
1376 | free_netdev(dev); |
1377 | dev_plip[i] = NULL; |
1378 | } |
1379 | } |
1380 | |
1381 | parport_unregister_driver(&plip_driver); |
1382 | } |
1383 | |
1384 | #ifndef MODULE |
1385 | |
1386 | static int parport_ptr; |
1387 | |
1388 | static int __init plip_setup(char *str) |
1389 | { |
1390 | int ints[4]; |
1391 | |
1392 | str = get_options(str, ARRAY_SIZE(ints), ints); |
1393 | |
1394 | /* Ugh. */ |
1395 | if (!strncmp(str, "parport" , 7)) { |
1396 | int n = simple_strtoul(str+7, NULL, 10); |
1397 | if (parport_ptr < PLIP_MAX) |
1398 | parport[parport_ptr++] = n; |
1399 | else |
1400 | printk(KERN_INFO "plip: too many ports, %s ignored.\n" , |
1401 | str); |
1402 | } else if (!strcmp(str, "timid" )) { |
1403 | timid = 1; |
1404 | } else { |
1405 | if (ints[0] == 0 || ints[1] == 0) { |
1406 | /* disable driver on "plip=" or "plip=0" */ |
1407 | parport[0] = -2; |
1408 | } else { |
1409 | printk(KERN_WARNING "warning: 'plip=0x%x' ignored\n" , |
1410 | ints[1]); |
1411 | } |
1412 | } |
1413 | return 1; |
1414 | } |
1415 | |
1416 | __setup("plip=" , plip_setup); |
1417 | |
1418 | #endif /* !MODULE */ |
1419 | |
1420 | static int __init plip_init (void) |
1421 | { |
1422 | if (parport[0] == -2) |
1423 | return 0; |
1424 | |
1425 | if (parport[0] != -1 && timid) { |
1426 | printk(KERN_WARNING "plip: warning, ignoring `timid' since specific ports given.\n" ); |
1427 | timid = 0; |
1428 | } |
1429 | |
1430 | if (parport_register_driver (&plip_driver)) { |
1431 | printk (KERN_WARNING "plip: couldn't register driver\n" ); |
1432 | return 1; |
1433 | } |
1434 | |
1435 | return 0; |
1436 | } |
1437 | |
1438 | module_init(plip_init); |
1439 | module_exit(plip_cleanup_module); |
1440 | MODULE_LICENSE("GPL" ); |
1441 | |