blob: f158fdf3aab2c931f8064f9ca419bd1729fafa68 [file] [log] [blame]
David Brazdil0f672f62019-12-10 10:32:29 +00001// SPDX-License-Identifier: GPL-2.0-or-later
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00002/*
3 * RDC R6040 Fast Ethernet MAC support
4 *
5 * Copyright (C) 2004 Sten Wang <sten.wang@rdc.com.tw>
6 * Copyright (C) 2007
7 * Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>
8 * Copyright (C) 2007-2012 Florian Fainelli <f.fainelli@gmail.com>
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00009*/
10
11#include <linux/kernel.h>
12#include <linux/module.h>
13#include <linux/moduleparam.h>
14#include <linux/string.h>
15#include <linux/timer.h>
16#include <linux/errno.h>
17#include <linux/ioport.h>
18#include <linux/interrupt.h>
19#include <linux/pci.h>
20#include <linux/netdevice.h>
21#include <linux/etherdevice.h>
22#include <linux/skbuff.h>
23#include <linux/delay.h>
24#include <linux/mii.h>
25#include <linux/ethtool.h>
26#include <linux/crc32.h>
27#include <linux/spinlock.h>
28#include <linux/bitops.h>
29#include <linux/io.h>
30#include <linux/irq.h>
31#include <linux/uaccess.h>
32#include <linux/phy.h>
33
34#include <asm/processor.h>
35
36#define DRV_NAME "r6040"
37#define DRV_VERSION "0.29"
38#define DRV_RELDATE "04Jul2016"
39
40/* Time in jiffies before concluding the transmitter is hung. */
41#define TX_TIMEOUT (6000 * HZ / 1000)
42
43/* RDC MAC I/O Size */
44#define R6040_IO_SIZE 256
45
46/* MAX RDC MAC */
47#define MAX_MAC 2
48
49/* MAC registers */
50#define MCR0 0x00 /* Control register 0 */
51#define MCR0_RCVEN 0x0002 /* Receive enable */
52#define MCR0_PROMISC 0x0020 /* Promiscuous mode */
53#define MCR0_HASH_EN 0x0100 /* Enable multicast hash table function */
54#define MCR0_XMTEN 0x1000 /* Transmission enable */
55#define MCR0_FD 0x8000 /* Full/Half duplex */
56#define MCR1 0x04 /* Control register 1 */
57#define MAC_RST 0x0001 /* Reset the MAC */
58#define MBCR 0x08 /* Bus control */
59#define MT_ICR 0x0C /* TX interrupt control */
60#define MR_ICR 0x10 /* RX interrupt control */
61#define MTPR 0x14 /* TX poll command register */
62#define TM2TX 0x0001 /* Trigger MAC to transmit */
63#define MR_BSR 0x18 /* RX buffer size */
64#define MR_DCR 0x1A /* RX descriptor control */
65#define MLSR 0x1C /* Last status */
66#define TX_FIFO_UNDR 0x0200 /* TX FIFO under-run */
67#define TX_EXCEEDC 0x2000 /* Transmit exceed collision */
68#define TX_LATEC 0x4000 /* Transmit late collision */
69#define MMDIO 0x20 /* MDIO control register */
70#define MDIO_WRITE 0x4000 /* MDIO write */
71#define MDIO_READ 0x2000 /* MDIO read */
72#define MMRD 0x24 /* MDIO read data register */
73#define MMWD 0x28 /* MDIO write data register */
74#define MTD_SA0 0x2C /* TX descriptor start address 0 */
75#define MTD_SA1 0x30 /* TX descriptor start address 1 */
76#define MRD_SA0 0x34 /* RX descriptor start address 0 */
77#define MRD_SA1 0x38 /* RX descriptor start address 1 */
78#define MISR 0x3C /* Status register */
79#define MIER 0x40 /* INT enable register */
80#define MSK_INT 0x0000 /* Mask off interrupts */
81#define RX_FINISH 0x0001 /* RX finished */
82#define RX_NO_DESC 0x0002 /* No RX descriptor available */
83#define RX_FIFO_FULL 0x0004 /* RX FIFO full */
84#define RX_EARLY 0x0008 /* RX early */
85#define TX_FINISH 0x0010 /* TX finished */
86#define TX_EARLY 0x0080 /* TX early */
87#define EVENT_OVRFL 0x0100 /* Event counter overflow */
88#define LINK_CHANGED 0x0200 /* PHY link changed */
89#define ME_CISR 0x44 /* Event counter INT status */
90#define ME_CIER 0x48 /* Event counter INT enable */
91#define MR_CNT 0x50 /* Successfully received packet counter */
92#define ME_CNT0 0x52 /* Event counter 0 */
93#define ME_CNT1 0x54 /* Event counter 1 */
94#define ME_CNT2 0x56 /* Event counter 2 */
95#define ME_CNT3 0x58 /* Event counter 3 */
96#define MT_CNT 0x5A /* Successfully transmit packet counter */
97#define ME_CNT4 0x5C /* Event counter 4 */
98#define MP_CNT 0x5E /* Pause frame counter register */
99#define MAR0 0x60 /* Hash table 0 */
100#define MAR1 0x62 /* Hash table 1 */
101#define MAR2 0x64 /* Hash table 2 */
102#define MAR3 0x66 /* Hash table 3 */
103#define MID_0L 0x68 /* Multicast address MID0 Low */
104#define MID_0M 0x6A /* Multicast address MID0 Medium */
105#define MID_0H 0x6C /* Multicast address MID0 High */
106#define MID_1L 0x70 /* MID1 Low */
107#define MID_1M 0x72 /* MID1 Medium */
108#define MID_1H 0x74 /* MID1 High */
109#define MID_2L 0x78 /* MID2 Low */
110#define MID_2M 0x7A /* MID2 Medium */
111#define MID_2H 0x7C /* MID2 High */
112#define MID_3L 0x80 /* MID3 Low */
113#define MID_3M 0x82 /* MID3 Medium */
114#define MID_3H 0x84 /* MID3 High */
115#define PHY_CC 0x88 /* PHY status change configuration register */
116#define SCEN 0x8000 /* PHY status change enable */
117#define PHYAD_SHIFT 8 /* PHY address shift */
118#define TMRDIV_SHIFT 0 /* Timer divider shift */
119#define PHY_ST 0x8A /* PHY status register */
120#define MAC_SM 0xAC /* MAC status machine */
121#define MAC_SM_RST 0x0002 /* MAC status machine reset */
Olivier Deprez0e641232021-09-23 10:07:05 +0200122#define MD_CSC 0xb6 /* MDC speed control register */
123#define MD_CSC_DEFAULT 0x0030
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000124#define MAC_ID 0xBE /* Identifier register */
125
126#define TX_DCNT 0x80 /* TX descriptor count */
127#define RX_DCNT 0x80 /* RX descriptor count */
128#define MAX_BUF_SIZE 0x600
129#define RX_DESC_SIZE (RX_DCNT * sizeof(struct r6040_descriptor))
130#define TX_DESC_SIZE (TX_DCNT * sizeof(struct r6040_descriptor))
131#define MBCR_DEFAULT 0x012A /* MAC Bus Control Register */
132#define MCAST_MAX 3 /* Max number multicast addresses to filter */
133
134#define MAC_DEF_TIMEOUT 2048 /* Default MAC read/write operation timeout */
135
136/* Descriptor status */
137#define DSC_OWNER_MAC 0x8000 /* MAC is the owner of this descriptor */
138#define DSC_RX_OK 0x4000 /* RX was successful */
139#define DSC_RX_ERR 0x0800 /* RX PHY error */
140#define DSC_RX_ERR_DRI 0x0400 /* RX dribble packet */
141#define DSC_RX_ERR_BUF 0x0200 /* RX length exceeds buffer size */
142#define DSC_RX_ERR_LONG 0x0100 /* RX length > maximum packet length */
143#define DSC_RX_ERR_RUNT 0x0080 /* RX packet length < 64 byte */
144#define DSC_RX_ERR_CRC 0x0040 /* RX CRC error */
145#define DSC_RX_BCAST 0x0020 /* RX broadcast (no error) */
146#define DSC_RX_MCAST 0x0010 /* RX multicast (no error) */
147#define DSC_RX_MCH_HIT 0x0008 /* RX multicast hit in hash table (no error) */
148#define DSC_RX_MIDH_HIT 0x0004 /* RX MID table hit (no error) */
149#define DSC_RX_IDX_MID_MASK 3 /* RX mask for the index of matched MIDx */
150
151MODULE_AUTHOR("Sten Wang <sten.wang@rdc.com.tw>,"
152 "Daniel Gimpelevich <daniel@gimpelevich.san-francisco.ca.us>,"
153 "Florian Fainelli <f.fainelli@gmail.com>");
154MODULE_LICENSE("GPL");
155MODULE_DESCRIPTION("RDC R6040 NAPI PCI FastEthernet driver");
156MODULE_VERSION(DRV_VERSION " " DRV_RELDATE);
157
158/* RX and TX interrupts that we handle */
159#define RX_INTS (RX_FIFO_FULL | RX_NO_DESC | RX_FINISH)
160#define TX_INTS (TX_FINISH)
161#define INT_MASK (RX_INTS | TX_INTS)
162
163struct r6040_descriptor {
164 u16 status, len; /* 0-3 */
165 __le32 buf; /* 4-7 */
166 __le32 ndesc; /* 8-B */
167 u32 rev1; /* C-F */
168 char *vbufp; /* 10-13 */
169 struct r6040_descriptor *vndescp; /* 14-17 */
170 struct sk_buff *skb_ptr; /* 18-1B */
171 u32 rev2; /* 1C-1F */
172} __aligned(32);
173
174struct r6040_private {
175 spinlock_t lock; /* driver lock */
176 struct pci_dev *pdev;
177 struct r6040_descriptor *rx_insert_ptr;
178 struct r6040_descriptor *rx_remove_ptr;
179 struct r6040_descriptor *tx_insert_ptr;
180 struct r6040_descriptor *tx_remove_ptr;
181 struct r6040_descriptor *rx_ring;
182 struct r6040_descriptor *tx_ring;
183 dma_addr_t rx_ring_dma;
184 dma_addr_t tx_ring_dma;
185 u16 tx_free_desc;
186 u16 mcr0;
187 struct net_device *dev;
188 struct mii_bus *mii_bus;
189 struct napi_struct napi;
190 void __iomem *base;
191 int old_link;
192 int old_duplex;
193};
194
195static char version[] = DRV_NAME
196 ": RDC R6040 NAPI net driver,"
197 "version "DRV_VERSION " (" DRV_RELDATE ")";
198
199/* Read a word data from PHY Chip */
200static int r6040_phy_read(void __iomem *ioaddr, int phy_addr, int reg)
201{
202 int limit = MAC_DEF_TIMEOUT;
203 u16 cmd;
204
205 iowrite16(MDIO_READ + reg + (phy_addr << 8), ioaddr + MMDIO);
206 /* Wait for the read bit to be cleared */
207 while (limit--) {
208 cmd = ioread16(ioaddr + MMDIO);
209 if (!(cmd & MDIO_READ))
210 break;
211 udelay(1);
212 }
213
214 if (limit < 0)
215 return -ETIMEDOUT;
216
217 return ioread16(ioaddr + MMRD);
218}
219
220/* Write a word data from PHY Chip */
221static int r6040_phy_write(void __iomem *ioaddr,
222 int phy_addr, int reg, u16 val)
223{
224 int limit = MAC_DEF_TIMEOUT;
225 u16 cmd;
226
227 iowrite16(val, ioaddr + MMWD);
228 /* Write the command to the MDIO bus */
229 iowrite16(MDIO_WRITE + reg + (phy_addr << 8), ioaddr + MMDIO);
230 /* Wait for the write bit to be cleared */
231 while (limit--) {
232 cmd = ioread16(ioaddr + MMDIO);
233 if (!(cmd & MDIO_WRITE))
234 break;
235 udelay(1);
236 }
237
238 return (limit < 0) ? -ETIMEDOUT : 0;
239}
240
241static int r6040_mdiobus_read(struct mii_bus *bus, int phy_addr, int reg)
242{
243 struct net_device *dev = bus->priv;
244 struct r6040_private *lp = netdev_priv(dev);
245 void __iomem *ioaddr = lp->base;
246
247 return r6040_phy_read(ioaddr, phy_addr, reg);
248}
249
250static int r6040_mdiobus_write(struct mii_bus *bus, int phy_addr,
251 int reg, u16 value)
252{
253 struct net_device *dev = bus->priv;
254 struct r6040_private *lp = netdev_priv(dev);
255 void __iomem *ioaddr = lp->base;
256
257 return r6040_phy_write(ioaddr, phy_addr, reg, value);
258}
259
260static void r6040_free_txbufs(struct net_device *dev)
261{
262 struct r6040_private *lp = netdev_priv(dev);
263 int i;
264
265 for (i = 0; i < TX_DCNT; i++) {
266 if (lp->tx_insert_ptr->skb_ptr) {
267 pci_unmap_single(lp->pdev,
268 le32_to_cpu(lp->tx_insert_ptr->buf),
269 MAX_BUF_SIZE, PCI_DMA_TODEVICE);
270 dev_kfree_skb(lp->tx_insert_ptr->skb_ptr);
271 lp->tx_insert_ptr->skb_ptr = NULL;
272 }
273 lp->tx_insert_ptr = lp->tx_insert_ptr->vndescp;
274 }
275}
276
277static void r6040_free_rxbufs(struct net_device *dev)
278{
279 struct r6040_private *lp = netdev_priv(dev);
280 int i;
281
282 for (i = 0; i < RX_DCNT; i++) {
283 if (lp->rx_insert_ptr->skb_ptr) {
284 pci_unmap_single(lp->pdev,
285 le32_to_cpu(lp->rx_insert_ptr->buf),
286 MAX_BUF_SIZE, PCI_DMA_FROMDEVICE);
287 dev_kfree_skb(lp->rx_insert_ptr->skb_ptr);
288 lp->rx_insert_ptr->skb_ptr = NULL;
289 }
290 lp->rx_insert_ptr = lp->rx_insert_ptr->vndescp;
291 }
292}
293
294static void r6040_init_ring_desc(struct r6040_descriptor *desc_ring,
295 dma_addr_t desc_dma, int size)
296{
297 struct r6040_descriptor *desc = desc_ring;
298 dma_addr_t mapping = desc_dma;
299
300 while (size-- > 0) {
301 mapping += sizeof(*desc);
302 desc->ndesc = cpu_to_le32(mapping);
303 desc->vndescp = desc + 1;
304 desc++;
305 }
306 desc--;
307 desc->ndesc = cpu_to_le32(desc_dma);
308 desc->vndescp = desc_ring;
309}
310
311static void r6040_init_txbufs(struct net_device *dev)
312{
313 struct r6040_private *lp = netdev_priv(dev);
314
315 lp->tx_free_desc = TX_DCNT;
316
317 lp->tx_remove_ptr = lp->tx_insert_ptr = lp->tx_ring;
318 r6040_init_ring_desc(lp->tx_ring, lp->tx_ring_dma, TX_DCNT);
319}
320
321static int r6040_alloc_rxbufs(struct net_device *dev)
322{
323 struct r6040_private *lp = netdev_priv(dev);
324 struct r6040_descriptor *desc;
325 struct sk_buff *skb;
326 int rc;
327
328 lp->rx_remove_ptr = lp->rx_insert_ptr = lp->rx_ring;
329 r6040_init_ring_desc(lp->rx_ring, lp->rx_ring_dma, RX_DCNT);
330
331 /* Allocate skbs for the rx descriptors */
332 desc = lp->rx_ring;
333 do {
334 skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
335 if (!skb) {
336 rc = -ENOMEM;
337 goto err_exit;
338 }
339 desc->skb_ptr = skb;
340 desc->buf = cpu_to_le32(pci_map_single(lp->pdev,
341 desc->skb_ptr->data,
342 MAX_BUF_SIZE, PCI_DMA_FROMDEVICE));
343 desc->status = DSC_OWNER_MAC;
344 desc = desc->vndescp;
345 } while (desc != lp->rx_ring);
346
347 return 0;
348
349err_exit:
350 /* Deallocate all previously allocated skbs */
351 r6040_free_rxbufs(dev);
352 return rc;
353}
354
355static void r6040_reset_mac(struct r6040_private *lp)
356{
357 void __iomem *ioaddr = lp->base;
358 int limit = MAC_DEF_TIMEOUT;
Olivier Deprez0e641232021-09-23 10:07:05 +0200359 u16 cmd, md_csc;
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000360
Olivier Deprez0e641232021-09-23 10:07:05 +0200361 md_csc = ioread16(ioaddr + MD_CSC);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000362 iowrite16(MAC_RST, ioaddr + MCR1);
363 while (limit--) {
364 cmd = ioread16(ioaddr + MCR1);
365 if (cmd & MAC_RST)
366 break;
367 }
368
369 /* Reset internal state machine */
370 iowrite16(MAC_SM_RST, ioaddr + MAC_SM);
371 iowrite16(0, ioaddr + MAC_SM);
372 mdelay(5);
Olivier Deprez0e641232021-09-23 10:07:05 +0200373
374 /* Restore MDIO clock frequency */
375 if (md_csc != MD_CSC_DEFAULT)
376 iowrite16(md_csc, ioaddr + MD_CSC);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000377}
378
379static void r6040_init_mac_regs(struct net_device *dev)
380{
381 struct r6040_private *lp = netdev_priv(dev);
382 void __iomem *ioaddr = lp->base;
383
384 /* Mask Off Interrupt */
385 iowrite16(MSK_INT, ioaddr + MIER);
386
387 /* Reset RDC MAC */
388 r6040_reset_mac(lp);
389
390 /* MAC Bus Control Register */
391 iowrite16(MBCR_DEFAULT, ioaddr + MBCR);
392
393 /* Buffer Size Register */
394 iowrite16(MAX_BUF_SIZE, ioaddr + MR_BSR);
395
396 /* Write TX ring start address */
397 iowrite16(lp->tx_ring_dma, ioaddr + MTD_SA0);
398 iowrite16(lp->tx_ring_dma >> 16, ioaddr + MTD_SA1);
399
400 /* Write RX ring start address */
401 iowrite16(lp->rx_ring_dma, ioaddr + MRD_SA0);
402 iowrite16(lp->rx_ring_dma >> 16, ioaddr + MRD_SA1);
403
404 /* Set interrupt waiting time and packet numbers */
405 iowrite16(0, ioaddr + MT_ICR);
406 iowrite16(0, ioaddr + MR_ICR);
407
408 /* Enable interrupts */
409 iowrite16(INT_MASK, ioaddr + MIER);
410
411 /* Enable TX and RX */
412 iowrite16(lp->mcr0 | MCR0_RCVEN, ioaddr);
413
414 /* Let TX poll the descriptors
415 * we may got called by r6040_tx_timeout which has left
416 * some unsent tx buffers */
417 iowrite16(TM2TX, ioaddr + MTPR);
418}
419
420static void r6040_tx_timeout(struct net_device *dev)
421{
422 struct r6040_private *priv = netdev_priv(dev);
423 void __iomem *ioaddr = priv->base;
424
425 netdev_warn(dev, "transmit timed out, int enable %4.4x "
426 "status %4.4x\n",
427 ioread16(ioaddr + MIER),
428 ioread16(ioaddr + MISR));
429
430 dev->stats.tx_errors++;
431
432 /* Reset MAC and re-init all registers */
433 r6040_init_mac_regs(dev);
434}
435
436static struct net_device_stats *r6040_get_stats(struct net_device *dev)
437{
438 struct r6040_private *priv = netdev_priv(dev);
439 void __iomem *ioaddr = priv->base;
440 unsigned long flags;
441
442 spin_lock_irqsave(&priv->lock, flags);
443 dev->stats.rx_crc_errors += ioread8(ioaddr + ME_CNT1);
444 dev->stats.multicast += ioread8(ioaddr + ME_CNT0);
445 spin_unlock_irqrestore(&priv->lock, flags);
446
447 return &dev->stats;
448}
449
450/* Stop RDC MAC and Free the allocated resource */
451static void r6040_down(struct net_device *dev)
452{
453 struct r6040_private *lp = netdev_priv(dev);
454 void __iomem *ioaddr = lp->base;
455 u16 *adrp;
456
457 /* Stop MAC */
458 iowrite16(MSK_INT, ioaddr + MIER); /* Mask Off Interrupt */
459
460 /* Reset RDC MAC */
461 r6040_reset_mac(lp);
462
463 /* Restore MAC Address to MIDx */
464 adrp = (u16 *) dev->dev_addr;
465 iowrite16(adrp[0], ioaddr + MID_0L);
466 iowrite16(adrp[1], ioaddr + MID_0M);
467 iowrite16(adrp[2], ioaddr + MID_0H);
468}
469
470static int r6040_close(struct net_device *dev)
471{
472 struct r6040_private *lp = netdev_priv(dev);
473 struct pci_dev *pdev = lp->pdev;
474
475 phy_stop(dev->phydev);
476 napi_disable(&lp->napi);
477 netif_stop_queue(dev);
478
479 spin_lock_irq(&lp->lock);
480 r6040_down(dev);
481
482 /* Free RX buffer */
483 r6040_free_rxbufs(dev);
484
485 /* Free TX buffer */
486 r6040_free_txbufs(dev);
487
488 spin_unlock_irq(&lp->lock);
489
490 free_irq(dev->irq, dev);
491
492 /* Free Descriptor memory */
493 if (lp->rx_ring) {
494 pci_free_consistent(pdev,
495 RX_DESC_SIZE, lp->rx_ring, lp->rx_ring_dma);
496 lp->rx_ring = NULL;
497 }
498
499 if (lp->tx_ring) {
500 pci_free_consistent(pdev,
501 TX_DESC_SIZE, lp->tx_ring, lp->tx_ring_dma);
502 lp->tx_ring = NULL;
503 }
504
505 return 0;
506}
507
508static int r6040_ioctl(struct net_device *dev, struct ifreq *rq, int cmd)
509{
510 if (!dev->phydev)
511 return -EINVAL;
512
513 return phy_mii_ioctl(dev->phydev, rq, cmd);
514}
515
516static int r6040_rx(struct net_device *dev, int limit)
517{
518 struct r6040_private *priv = netdev_priv(dev);
519 struct r6040_descriptor *descptr = priv->rx_remove_ptr;
520 struct sk_buff *skb_ptr, *new_skb;
521 int count = 0;
522 u16 err;
523
524 /* Limit not reached and the descriptor belongs to the CPU */
525 while (count < limit && !(descptr->status & DSC_OWNER_MAC)) {
526 /* Read the descriptor status */
527 err = descptr->status;
528 /* Global error status set */
529 if (err & DSC_RX_ERR) {
530 /* RX dribble */
531 if (err & DSC_RX_ERR_DRI)
532 dev->stats.rx_frame_errors++;
533 /* Buffer length exceeded */
534 if (err & DSC_RX_ERR_BUF)
535 dev->stats.rx_length_errors++;
536 /* Packet too long */
537 if (err & DSC_RX_ERR_LONG)
538 dev->stats.rx_length_errors++;
539 /* Packet < 64 bytes */
540 if (err & DSC_RX_ERR_RUNT)
541 dev->stats.rx_length_errors++;
542 /* CRC error */
543 if (err & DSC_RX_ERR_CRC) {
544 spin_lock(&priv->lock);
545 dev->stats.rx_crc_errors++;
546 spin_unlock(&priv->lock);
547 }
548 goto next_descr;
549 }
550
551 /* Packet successfully received */
552 new_skb = netdev_alloc_skb(dev, MAX_BUF_SIZE);
553 if (!new_skb) {
554 dev->stats.rx_dropped++;
555 goto next_descr;
556 }
557 skb_ptr = descptr->skb_ptr;
558 skb_ptr->dev = priv->dev;
559
560 /* Do not count the CRC */
561 skb_put(skb_ptr, descptr->len - 4);
562 pci_unmap_single(priv->pdev, le32_to_cpu(descptr->buf),
563 MAX_BUF_SIZE, PCI_DMA_FROMDEVICE);
564 skb_ptr->protocol = eth_type_trans(skb_ptr, priv->dev);
565
566 /* Send to upper layer */
567 netif_receive_skb(skb_ptr);
568 dev->stats.rx_packets++;
569 dev->stats.rx_bytes += descptr->len - 4;
570
571 /* put new skb into descriptor */
572 descptr->skb_ptr = new_skb;
573 descptr->buf = cpu_to_le32(pci_map_single(priv->pdev,
574 descptr->skb_ptr->data,
575 MAX_BUF_SIZE, PCI_DMA_FROMDEVICE));
576
577next_descr:
578 /* put the descriptor back to the MAC */
579 descptr->status = DSC_OWNER_MAC;
580 descptr = descptr->vndescp;
581 count++;
582 }
583 priv->rx_remove_ptr = descptr;
584
585 return count;
586}
587
588static void r6040_tx(struct net_device *dev)
589{
590 struct r6040_private *priv = netdev_priv(dev);
591 struct r6040_descriptor *descptr;
592 void __iomem *ioaddr = priv->base;
593 struct sk_buff *skb_ptr;
594 u16 err;
595
596 spin_lock(&priv->lock);
597 descptr = priv->tx_remove_ptr;
598 while (priv->tx_free_desc < TX_DCNT) {
599 /* Check for errors */
600 err = ioread16(ioaddr + MLSR);
601
602 if (err & TX_FIFO_UNDR)
603 dev->stats.tx_fifo_errors++;
604 if (err & (TX_EXCEEDC | TX_LATEC))
605 dev->stats.tx_carrier_errors++;
606
607 if (descptr->status & DSC_OWNER_MAC)
608 break; /* Not complete */
609 skb_ptr = descptr->skb_ptr;
610
611 /* Statistic Counter */
612 dev->stats.tx_packets++;
613 dev->stats.tx_bytes += skb_ptr->len;
614
615 pci_unmap_single(priv->pdev, le32_to_cpu(descptr->buf),
616 skb_ptr->len, PCI_DMA_TODEVICE);
617 /* Free buffer */
618 dev_kfree_skb(skb_ptr);
619 descptr->skb_ptr = NULL;
620 /* To next descriptor */
621 descptr = descptr->vndescp;
622 priv->tx_free_desc++;
623 }
624 priv->tx_remove_ptr = descptr;
625
626 if (priv->tx_free_desc)
627 netif_wake_queue(dev);
628 spin_unlock(&priv->lock);
629}
630
631static int r6040_poll(struct napi_struct *napi, int budget)
632{
633 struct r6040_private *priv =
634 container_of(napi, struct r6040_private, napi);
635 struct net_device *dev = priv->dev;
636 void __iomem *ioaddr = priv->base;
637 int work_done;
638
639 r6040_tx(dev);
640
641 work_done = r6040_rx(dev, budget);
642
643 if (work_done < budget) {
644 napi_complete_done(napi, work_done);
645 /* Enable RX/TX interrupt */
646 iowrite16(ioread16(ioaddr + MIER) | RX_INTS | TX_INTS,
647 ioaddr + MIER);
648 }
649 return work_done;
650}
651
652/* The RDC interrupt handler. */
653static irqreturn_t r6040_interrupt(int irq, void *dev_id)
654{
655 struct net_device *dev = dev_id;
656 struct r6040_private *lp = netdev_priv(dev);
657 void __iomem *ioaddr = lp->base;
658 u16 misr, status;
659
660 /* Save MIER */
661 misr = ioread16(ioaddr + MIER);
662 /* Mask off RDC MAC interrupt */
663 iowrite16(MSK_INT, ioaddr + MIER);
664 /* Read MISR status and clear */
665 status = ioread16(ioaddr + MISR);
666
667 if (status == 0x0000 || status == 0xffff) {
668 /* Restore RDC MAC interrupt */
669 iowrite16(misr, ioaddr + MIER);
670 return IRQ_NONE;
671 }
672
673 /* RX interrupt request */
674 if (status & (RX_INTS | TX_INTS)) {
675 if (status & RX_NO_DESC) {
676 /* RX descriptor unavailable */
677 dev->stats.rx_dropped++;
678 dev->stats.rx_missed_errors++;
679 }
680 if (status & RX_FIFO_FULL)
681 dev->stats.rx_fifo_errors++;
682
683 if (likely(napi_schedule_prep(&lp->napi))) {
684 /* Mask off RX interrupt */
685 misr &= ~(RX_INTS | TX_INTS);
686 __napi_schedule_irqoff(&lp->napi);
687 }
688 }
689
690 /* Restore RDC MAC interrupt */
691 iowrite16(misr, ioaddr + MIER);
692
693 return IRQ_HANDLED;
694}
695
696#ifdef CONFIG_NET_POLL_CONTROLLER
697static void r6040_poll_controller(struct net_device *dev)
698{
699 disable_irq(dev->irq);
700 r6040_interrupt(dev->irq, dev);
701 enable_irq(dev->irq);
702}
703#endif
704
705/* Init RDC MAC */
706static int r6040_up(struct net_device *dev)
707{
708 struct r6040_private *lp = netdev_priv(dev);
709 void __iomem *ioaddr = lp->base;
710 int ret;
711
712 /* Initialise and alloc RX/TX buffers */
713 r6040_init_txbufs(dev);
714 ret = r6040_alloc_rxbufs(dev);
715 if (ret)
716 return ret;
717
718 /* improve performance (by RDC guys) */
719 r6040_phy_write(ioaddr, 30, 17,
720 (r6040_phy_read(ioaddr, 30, 17) | 0x4000));
721 r6040_phy_write(ioaddr, 30, 17,
722 ~((~r6040_phy_read(ioaddr, 30, 17)) | 0x2000));
723 r6040_phy_write(ioaddr, 0, 19, 0x0000);
724 r6040_phy_write(ioaddr, 0, 30, 0x01F0);
725
726 /* Initialize all MAC registers */
727 r6040_init_mac_regs(dev);
728
729 phy_start(dev->phydev);
730
731 return 0;
732}
733
734
735/* Read/set MAC address routines */
736static void r6040_mac_address(struct net_device *dev)
737{
738 struct r6040_private *lp = netdev_priv(dev);
739 void __iomem *ioaddr = lp->base;
740 u16 *adrp;
741
742 /* Reset MAC */
743 r6040_reset_mac(lp);
744
745 /* Restore MAC Address */
746 adrp = (u16 *) dev->dev_addr;
747 iowrite16(adrp[0], ioaddr + MID_0L);
748 iowrite16(adrp[1], ioaddr + MID_0M);
749 iowrite16(adrp[2], ioaddr + MID_0H);
750}
751
752static int r6040_open(struct net_device *dev)
753{
754 struct r6040_private *lp = netdev_priv(dev);
755 int ret;
756
757 /* Request IRQ and Register interrupt handler */
758 ret = request_irq(dev->irq, r6040_interrupt,
759 IRQF_SHARED, dev->name, dev);
760 if (ret)
761 goto out;
762
763 /* Set MAC address */
764 r6040_mac_address(dev);
765
766 /* Allocate Descriptor memory */
767 lp->rx_ring =
768 pci_alloc_consistent(lp->pdev, RX_DESC_SIZE, &lp->rx_ring_dma);
769 if (!lp->rx_ring) {
770 ret = -ENOMEM;
771 goto err_free_irq;
772 }
773
774 lp->tx_ring =
775 pci_alloc_consistent(lp->pdev, TX_DESC_SIZE, &lp->tx_ring_dma);
776 if (!lp->tx_ring) {
777 ret = -ENOMEM;
778 goto err_free_rx_ring;
779 }
780
781 ret = r6040_up(dev);
782 if (ret)
783 goto err_free_tx_ring;
784
785 napi_enable(&lp->napi);
786 netif_start_queue(dev);
787
788 return 0;
789
790err_free_tx_ring:
791 pci_free_consistent(lp->pdev, TX_DESC_SIZE, lp->tx_ring,
792 lp->tx_ring_dma);
793err_free_rx_ring:
794 pci_free_consistent(lp->pdev, RX_DESC_SIZE, lp->rx_ring,
795 lp->rx_ring_dma);
796err_free_irq:
797 free_irq(dev->irq, dev);
798out:
799 return ret;
800}
801
802static netdev_tx_t r6040_start_xmit(struct sk_buff *skb,
803 struct net_device *dev)
804{
805 struct r6040_private *lp = netdev_priv(dev);
806 struct r6040_descriptor *descptr;
807 void __iomem *ioaddr = lp->base;
808 unsigned long flags;
809
810 if (skb_put_padto(skb, ETH_ZLEN) < 0)
811 return NETDEV_TX_OK;
812
813 /* Critical Section */
814 spin_lock_irqsave(&lp->lock, flags);
815
816 /* TX resource check */
817 if (!lp->tx_free_desc) {
818 spin_unlock_irqrestore(&lp->lock, flags);
819 netif_stop_queue(dev);
820 netdev_err(dev, ": no tx descriptor\n");
821 return NETDEV_TX_BUSY;
822 }
823
824 /* Set TX descriptor & Transmit it */
825 lp->tx_free_desc--;
826 descptr = lp->tx_insert_ptr;
827 descptr->len = skb->len;
828 descptr->skb_ptr = skb;
829 descptr->buf = cpu_to_le32(pci_map_single(lp->pdev,
830 skb->data, skb->len, PCI_DMA_TODEVICE));
831 descptr->status = DSC_OWNER_MAC;
832
833 skb_tx_timestamp(skb);
834
835 /* Trigger the MAC to check the TX descriptor */
David Brazdil0f672f62019-12-10 10:32:29 +0000836 if (!netdev_xmit_more() || netif_queue_stopped(dev))
Andrew Scullb4b6d4a2019-01-02 15:54:55 +0000837 iowrite16(TM2TX, ioaddr + MTPR);
838 lp->tx_insert_ptr = descptr->vndescp;
839
840 /* If no tx resource, stop */
841 if (!lp->tx_free_desc)
842 netif_stop_queue(dev);
843
844 spin_unlock_irqrestore(&lp->lock, flags);
845
846 return NETDEV_TX_OK;
847}
848
849static void r6040_multicast_list(struct net_device *dev)
850{
851 struct r6040_private *lp = netdev_priv(dev);
852 void __iomem *ioaddr = lp->base;
853 unsigned long flags;
854 struct netdev_hw_addr *ha;
855 int i;
856 u16 *adrp;
857 u16 hash_table[4] = { 0 };
858
859 spin_lock_irqsave(&lp->lock, flags);
860
861 /* Keep our MAC Address */
862 adrp = (u16 *)dev->dev_addr;
863 iowrite16(adrp[0], ioaddr + MID_0L);
864 iowrite16(adrp[1], ioaddr + MID_0M);
865 iowrite16(adrp[2], ioaddr + MID_0H);
866
867 /* Clear AMCP & PROM bits */
868 lp->mcr0 = ioread16(ioaddr + MCR0) & ~(MCR0_PROMISC | MCR0_HASH_EN);
869
870 /* Promiscuous mode */
871 if (dev->flags & IFF_PROMISC)
872 lp->mcr0 |= MCR0_PROMISC;
873
874 /* Enable multicast hash table function to
875 * receive all multicast packets. */
876 else if (dev->flags & IFF_ALLMULTI) {
877 lp->mcr0 |= MCR0_HASH_EN;
878
879 for (i = 0; i < MCAST_MAX ; i++) {
880 iowrite16(0, ioaddr + MID_1L + 8 * i);
881 iowrite16(0, ioaddr + MID_1M + 8 * i);
882 iowrite16(0, ioaddr + MID_1H + 8 * i);
883 }
884
885 for (i = 0; i < 4; i++)
886 hash_table[i] = 0xffff;
887 }
888 /* Use internal multicast address registers if the number of
889 * multicast addresses is not greater than MCAST_MAX. */
890 else if (netdev_mc_count(dev) <= MCAST_MAX) {
891 i = 0;
892 netdev_for_each_mc_addr(ha, dev) {
893 u16 *adrp = (u16 *) ha->addr;
894 iowrite16(adrp[0], ioaddr + MID_1L + 8 * i);
895 iowrite16(adrp[1], ioaddr + MID_1M + 8 * i);
896 iowrite16(adrp[2], ioaddr + MID_1H + 8 * i);
897 i++;
898 }
899 while (i < MCAST_MAX) {
900 iowrite16(0, ioaddr + MID_1L + 8 * i);
901 iowrite16(0, ioaddr + MID_1M + 8 * i);
902 iowrite16(0, ioaddr + MID_1H + 8 * i);
903 i++;
904 }
905 }
906 /* Otherwise, Enable multicast hash table function. */
907 else {
908 u32 crc;
909
910 lp->mcr0 |= MCR0_HASH_EN;
911
912 for (i = 0; i < MCAST_MAX ; i++) {
913 iowrite16(0, ioaddr + MID_1L + 8 * i);
914 iowrite16(0, ioaddr + MID_1M + 8 * i);
915 iowrite16(0, ioaddr + MID_1H + 8 * i);
916 }
917
918 /* Build multicast hash table */
919 netdev_for_each_mc_addr(ha, dev) {
920 u8 *addrs = ha->addr;
921
922 crc = ether_crc(ETH_ALEN, addrs);
923 crc >>= 26;
924 hash_table[crc >> 4] |= 1 << (crc & 0xf);
925 }
926 }
927
928 iowrite16(lp->mcr0, ioaddr + MCR0);
929
930 /* Fill the MAC hash tables with their values */
931 if (lp->mcr0 & MCR0_HASH_EN) {
932 iowrite16(hash_table[0], ioaddr + MAR0);
933 iowrite16(hash_table[1], ioaddr + MAR1);
934 iowrite16(hash_table[2], ioaddr + MAR2);
935 iowrite16(hash_table[3], ioaddr + MAR3);
936 }
937
938 spin_unlock_irqrestore(&lp->lock, flags);
939}
940
941static void netdev_get_drvinfo(struct net_device *dev,
942 struct ethtool_drvinfo *info)
943{
944 struct r6040_private *rp = netdev_priv(dev);
945
946 strlcpy(info->driver, DRV_NAME, sizeof(info->driver));
947 strlcpy(info->version, DRV_VERSION, sizeof(info->version));
948 strlcpy(info->bus_info, pci_name(rp->pdev), sizeof(info->bus_info));
949}
950
951static const struct ethtool_ops netdev_ethtool_ops = {
952 .get_drvinfo = netdev_get_drvinfo,
953 .get_link = ethtool_op_get_link,
954 .get_ts_info = ethtool_op_get_ts_info,
955 .get_link_ksettings = phy_ethtool_get_link_ksettings,
956 .set_link_ksettings = phy_ethtool_set_link_ksettings,
957};
958
959static const struct net_device_ops r6040_netdev_ops = {
960 .ndo_open = r6040_open,
961 .ndo_stop = r6040_close,
962 .ndo_start_xmit = r6040_start_xmit,
963 .ndo_get_stats = r6040_get_stats,
964 .ndo_set_rx_mode = r6040_multicast_list,
965 .ndo_validate_addr = eth_validate_addr,
966 .ndo_set_mac_address = eth_mac_addr,
967 .ndo_do_ioctl = r6040_ioctl,
968 .ndo_tx_timeout = r6040_tx_timeout,
969#ifdef CONFIG_NET_POLL_CONTROLLER
970 .ndo_poll_controller = r6040_poll_controller,
971#endif
972};
973
974static void r6040_adjust_link(struct net_device *dev)
975{
976 struct r6040_private *lp = netdev_priv(dev);
977 struct phy_device *phydev = dev->phydev;
978 int status_changed = 0;
979 void __iomem *ioaddr = lp->base;
980
981 BUG_ON(!phydev);
982
983 if (lp->old_link != phydev->link) {
984 status_changed = 1;
985 lp->old_link = phydev->link;
986 }
987
988 /* reflect duplex change */
989 if (phydev->link && (lp->old_duplex != phydev->duplex)) {
990 lp->mcr0 |= (phydev->duplex == DUPLEX_FULL ? MCR0_FD : 0);
991 iowrite16(lp->mcr0, ioaddr);
992
993 status_changed = 1;
994 lp->old_duplex = phydev->duplex;
995 }
996
997 if (status_changed)
998 phy_print_status(phydev);
999}
1000
1001static int r6040_mii_probe(struct net_device *dev)
1002{
1003 struct r6040_private *lp = netdev_priv(dev);
1004 struct phy_device *phydev = NULL;
1005
1006 phydev = phy_find_first(lp->mii_bus);
1007 if (!phydev) {
1008 dev_err(&lp->pdev->dev, "no PHY found\n");
1009 return -ENODEV;
1010 }
1011
1012 phydev = phy_connect(dev, phydev_name(phydev), &r6040_adjust_link,
1013 PHY_INTERFACE_MODE_MII);
1014
1015 if (IS_ERR(phydev)) {
1016 dev_err(&lp->pdev->dev, "could not attach to PHY\n");
1017 return PTR_ERR(phydev);
1018 }
1019
David Brazdil0f672f62019-12-10 10:32:29 +00001020 phy_set_max_speed(phydev, SPEED_100);
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001021
Andrew Scullb4b6d4a2019-01-02 15:54:55 +00001022 lp->old_link = 0;
1023 lp->old_duplex = -1;
1024
1025 phy_attached_info(phydev);
1026
1027 return 0;
1028}
1029
1030static int r6040_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
1031{
1032 struct net_device *dev;
1033 struct r6040_private *lp;
1034 void __iomem *ioaddr;
1035 int err, io_size = R6040_IO_SIZE;
1036 static int card_idx = -1;
1037 int bar = 0;
1038 u16 *adrp;
1039
1040 pr_info("%s\n", version);
1041
1042 err = pci_enable_device(pdev);
1043 if (err)
1044 goto err_out;
1045
1046 /* this should always be supported */
1047 err = pci_set_dma_mask(pdev, DMA_BIT_MASK(32));
1048 if (err) {
1049 dev_err(&pdev->dev, "32-bit PCI DMA addresses not supported by the card\n");
1050 goto err_out_disable_dev;
1051 }
1052 err = pci_set_consistent_dma_mask(pdev, DMA_BIT_MASK(32));
1053 if (err) {
1054 dev_err(&pdev->dev, "32-bit PCI DMA addresses not supported by the card\n");
1055 goto err_out_disable_dev;
1056 }
1057
1058 /* IO Size check */
1059 if (pci_resource_len(pdev, bar) < io_size) {
1060 dev_err(&pdev->dev, "Insufficient PCI resources, aborting\n");
1061 err = -EIO;
1062 goto err_out_disable_dev;
1063 }
1064
1065 pci_set_master(pdev);
1066
1067 dev = alloc_etherdev(sizeof(struct r6040_private));
1068 if (!dev) {
1069 err = -ENOMEM;
1070 goto err_out_disable_dev;
1071 }
1072 SET_NETDEV_DEV(dev, &pdev->dev);
1073 lp = netdev_priv(dev);
1074
1075 err = pci_request_regions(pdev, DRV_NAME);
1076
1077 if (err) {
1078 dev_err(&pdev->dev, "Failed to request PCI regions\n");
1079 goto err_out_free_dev;
1080 }
1081
1082 ioaddr = pci_iomap(pdev, bar, io_size);
1083 if (!ioaddr) {
1084 dev_err(&pdev->dev, "ioremap failed for device\n");
1085 err = -EIO;
1086 goto err_out_free_res;
1087 }
1088
1089 /* If PHY status change register is still set to zero it means the
1090 * bootloader didn't initialize it, so we set it to:
1091 * - enable phy status change
1092 * - enable all phy addresses
1093 * - set to lowest timer divider */
1094 if (ioread16(ioaddr + PHY_CC) == 0)
1095 iowrite16(SCEN | PHY_MAX_ADDR << PHYAD_SHIFT |
1096 7 << TMRDIV_SHIFT, ioaddr + PHY_CC);
1097
1098 /* Init system & device */
1099 lp->base = ioaddr;
1100 dev->irq = pdev->irq;
1101
1102 spin_lock_init(&lp->lock);
1103 pci_set_drvdata(pdev, dev);
1104
1105 /* Set MAC address */
1106 card_idx++;
1107
1108 adrp = (u16 *)dev->dev_addr;
1109 adrp[0] = ioread16(ioaddr + MID_0L);
1110 adrp[1] = ioread16(ioaddr + MID_0M);
1111 adrp[2] = ioread16(ioaddr + MID_0H);
1112
1113 /* Some bootloader/BIOSes do not initialize
1114 * MAC address, warn about that */
1115 if (!(adrp[0] || adrp[1] || adrp[2])) {
1116 netdev_warn(dev, "MAC address not initialized, "
1117 "generating random\n");
1118 eth_hw_addr_random(dev);
1119 }
1120
1121 /* Link new device into r6040_root_dev */
1122 lp->pdev = pdev;
1123 lp->dev = dev;
1124
1125 /* Init RDC private data */
1126 lp->mcr0 = MCR0_XMTEN | MCR0_RCVEN;
1127
1128 /* The RDC-specific entries in the device structure. */
1129 dev->netdev_ops = &r6040_netdev_ops;
1130 dev->ethtool_ops = &netdev_ethtool_ops;
1131 dev->watchdog_timeo = TX_TIMEOUT;
1132
1133 netif_napi_add(dev, &lp->napi, r6040_poll, 64);
1134
1135 lp->mii_bus = mdiobus_alloc();
1136 if (!lp->mii_bus) {
1137 dev_err(&pdev->dev, "mdiobus_alloc() failed\n");
1138 err = -ENOMEM;
1139 goto err_out_unmap;
1140 }
1141
1142 lp->mii_bus->priv = dev;
1143 lp->mii_bus->read = r6040_mdiobus_read;
1144 lp->mii_bus->write = r6040_mdiobus_write;
1145 lp->mii_bus->name = "r6040_eth_mii";
1146 snprintf(lp->mii_bus->id, MII_BUS_ID_SIZE, "%s-%x",
1147 dev_name(&pdev->dev), card_idx);
1148
1149 err = mdiobus_register(lp->mii_bus);
1150 if (err) {
1151 dev_err(&pdev->dev, "failed to register MII bus\n");
1152 goto err_out_mdio;
1153 }
1154
1155 err = r6040_mii_probe(dev);
1156 if (err) {
1157 dev_err(&pdev->dev, "failed to probe MII bus\n");
1158 goto err_out_mdio_unregister;
1159 }
1160
1161 /* Register net device. After this dev->name assign */
1162 err = register_netdev(dev);
1163 if (err) {
1164 dev_err(&pdev->dev, "Failed to register net device\n");
1165 goto err_out_mdio_unregister;
1166 }
1167 return 0;
1168
1169err_out_mdio_unregister:
1170 mdiobus_unregister(lp->mii_bus);
1171err_out_mdio:
1172 mdiobus_free(lp->mii_bus);
1173err_out_unmap:
1174 netif_napi_del(&lp->napi);
1175 pci_iounmap(pdev, ioaddr);
1176err_out_free_res:
1177 pci_release_regions(pdev);
1178err_out_free_dev:
1179 free_netdev(dev);
1180err_out_disable_dev:
1181 pci_disable_device(pdev);
1182err_out:
1183 return err;
1184}
1185
1186static void r6040_remove_one(struct pci_dev *pdev)
1187{
1188 struct net_device *dev = pci_get_drvdata(pdev);
1189 struct r6040_private *lp = netdev_priv(dev);
1190
1191 unregister_netdev(dev);
1192 mdiobus_unregister(lp->mii_bus);
1193 mdiobus_free(lp->mii_bus);
1194 netif_napi_del(&lp->napi);
1195 pci_iounmap(pdev, lp->base);
1196 pci_release_regions(pdev);
1197 free_netdev(dev);
1198 pci_disable_device(pdev);
1199}
1200
1201
1202static const struct pci_device_id r6040_pci_tbl[] = {
1203 { PCI_DEVICE(PCI_VENDOR_ID_RDC, 0x6040) },
1204 { 0 }
1205};
1206MODULE_DEVICE_TABLE(pci, r6040_pci_tbl);
1207
1208static struct pci_driver r6040_driver = {
1209 .name = DRV_NAME,
1210 .id_table = r6040_pci_tbl,
1211 .probe = r6040_init_one,
1212 .remove = r6040_remove_one,
1213};
1214
1215module_pci_driver(r6040_driver);