David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 1 | // SPDX-License-Identifier: GPL-2.0-or-later |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 2 | |
| 3 | /* Advanced Micro Devices Inc. AMD8111E Linux Network Driver |
| 4 | * Copyright (C) 2004 Advanced Micro Devices |
| 5 | * |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 6 | * Copyright 2001,2002 Jeff Garzik <jgarzik@mandrakesoft.com> [ 8139cp.c,tg3.c ] |
| 7 | * Copyright (C) 2001, 2002 David S. Miller (davem@redhat.com)[ tg3.c] |
| 8 | * Copyright 1996-1999 Thomas Bogendoerfer [ pcnet32.c ] |
| 9 | * Derived from the lance driver written 1993,1994,1995 by Donald Becker. |
| 10 | * Copyright 1993 United States Government as represented by the |
| 11 | * Director, National Security Agency.[ pcnet32.c ] |
| 12 | * Carsten Langgaard, carstenl@mips.com [ pcnet32.c ] |
| 13 | * Copyright (C) 2000 MIPS Technologies, Inc. All rights reserved. |
| 14 | * |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 15 | |
| 16 | Module Name: |
| 17 | |
| 18 | amd8111e.c |
| 19 | |
| 20 | Abstract: |
| 21 | |
| 22 | AMD8111 based 10/100 Ethernet Controller Driver. |
| 23 | |
| 24 | Environment: |
| 25 | |
| 26 | Kernel Mode |
| 27 | |
| 28 | Revision History: |
| 29 | 3.0.0 |
| 30 | Initial Revision. |
| 31 | 3.0.1 |
| 32 | 1. Dynamic interrupt coalescing. |
| 33 | 2. Removed prev_stats. |
| 34 | 3. MII support. |
| 35 | 4. Dynamic IPG support |
| 36 | 3.0.2 05/29/2003 |
| 37 | 1. Bug fix: Fixed failure to send jumbo packets larger than 4k. |
| 38 | 2. Bug fix: Fixed VLAN support failure. |
| 39 | 3. Bug fix: Fixed receive interrupt coalescing bug. |
| 40 | 4. Dynamic IPG support is disabled by default. |
| 41 | 3.0.3 06/05/2003 |
| 42 | 1. Bug fix: Fixed failure to close the interface if SMP is enabled. |
| 43 | 3.0.4 12/09/2003 |
| 44 | 1. Added set_mac_address routine for bonding driver support. |
| 45 | 2. Tested the driver for bonding support |
| 46 | 3. Bug fix: Fixed mismach in actual receive buffer lenth and lenth |
| 47 | indicated to the h/w. |
| 48 | 4. Modified amd8111e_rx() routine to receive all the received packets |
| 49 | in the first interrupt. |
| 50 | 5. Bug fix: Corrected rx_errors reported in get_stats() function. |
| 51 | 3.0.5 03/22/2004 |
| 52 | 1. Added NAPI support |
| 53 | |
| 54 | */ |
| 55 | |
| 56 | |
| 57 | #include <linux/module.h> |
| 58 | #include <linux/kernel.h> |
| 59 | #include <linux/types.h> |
| 60 | #include <linux/compiler.h> |
| 61 | #include <linux/delay.h> |
| 62 | #include <linux/interrupt.h> |
| 63 | #include <linux/ioport.h> |
| 64 | #include <linux/pci.h> |
| 65 | #include <linux/netdevice.h> |
| 66 | #include <linux/etherdevice.h> |
| 67 | #include <linux/skbuff.h> |
| 68 | #include <linux/ethtool.h> |
| 69 | #include <linux/mii.h> |
| 70 | #include <linux/if_vlan.h> |
| 71 | #include <linux/ctype.h> |
| 72 | #include <linux/crc32.h> |
| 73 | #include <linux/dma-mapping.h> |
| 74 | |
| 75 | #include <asm/io.h> |
| 76 | #include <asm/byteorder.h> |
| 77 | #include <linux/uaccess.h> |
| 78 | |
| 79 | #if IS_ENABLED(CONFIG_VLAN_8021Q) |
| 80 | #define AMD8111E_VLAN_TAG_USED 1 |
| 81 | #else |
| 82 | #define AMD8111E_VLAN_TAG_USED 0 |
| 83 | #endif |
| 84 | |
| 85 | #include "amd8111e.h" |
| 86 | #define MODULE_NAME "amd8111e" |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 87 | MODULE_AUTHOR("Advanced Micro Devices, Inc."); |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 88 | MODULE_DESCRIPTION("AMD8111 based 10/100 Ethernet Controller."); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 89 | MODULE_LICENSE("GPL"); |
| 90 | module_param_array(speed_duplex, int, NULL, 0); |
| 91 | MODULE_PARM_DESC(speed_duplex, "Set device speed and duplex modes, 0: Auto Negotiate, 1: 10Mbps Half Duplex, 2: 10Mbps Full Duplex, 3: 100Mbps Half Duplex, 4: 100Mbps Full Duplex"); |
| 92 | module_param_array(coalesce, bool, NULL, 0); |
| 93 | MODULE_PARM_DESC(coalesce, "Enable or Disable interrupt coalescing, 1: Enable, 0: Disable"); |
| 94 | module_param_array(dynamic_ipg, bool, NULL, 0); |
| 95 | MODULE_PARM_DESC(dynamic_ipg, "Enable or Disable dynamic IPG, 1: Enable, 0: Disable"); |
| 96 | |
| 97 | /* This function will read the PHY registers. */ |
| 98 | static int amd8111e_read_phy(struct amd8111e_priv *lp, |
| 99 | int phy_id, int reg, u32 *val) |
| 100 | { |
| 101 | void __iomem *mmio = lp->mmio; |
| 102 | unsigned int reg_val; |
| 103 | unsigned int repeat= REPEAT_CNT; |
| 104 | |
| 105 | reg_val = readl(mmio + PHY_ACCESS); |
| 106 | while (reg_val & PHY_CMD_ACTIVE) |
| 107 | reg_val = readl( mmio + PHY_ACCESS ); |
| 108 | |
| 109 | writel( PHY_RD_CMD | ((phy_id & 0x1f) << 21) | |
| 110 | ((reg & 0x1f) << 16), mmio +PHY_ACCESS); |
| 111 | do{ |
| 112 | reg_val = readl(mmio + PHY_ACCESS); |
| 113 | udelay(30); /* It takes 30 us to read/write data */ |
| 114 | } while (--repeat && (reg_val & PHY_CMD_ACTIVE)); |
| 115 | if(reg_val & PHY_RD_ERR) |
| 116 | goto err_phy_read; |
| 117 | |
| 118 | *val = reg_val & 0xffff; |
| 119 | return 0; |
| 120 | err_phy_read: |
| 121 | *val = 0; |
| 122 | return -EINVAL; |
| 123 | |
| 124 | } |
| 125 | |
| 126 | /* This function will write into PHY registers. */ |
| 127 | static int amd8111e_write_phy(struct amd8111e_priv *lp, |
| 128 | int phy_id, int reg, u32 val) |
| 129 | { |
| 130 | unsigned int repeat = REPEAT_CNT; |
| 131 | void __iomem *mmio = lp->mmio; |
| 132 | unsigned int reg_val; |
| 133 | |
| 134 | reg_val = readl(mmio + PHY_ACCESS); |
| 135 | while (reg_val & PHY_CMD_ACTIVE) |
| 136 | reg_val = readl( mmio + PHY_ACCESS ); |
| 137 | |
| 138 | writel( PHY_WR_CMD | ((phy_id & 0x1f) << 21) | |
| 139 | ((reg & 0x1f) << 16)|val, mmio + PHY_ACCESS); |
| 140 | |
| 141 | do{ |
| 142 | reg_val = readl(mmio + PHY_ACCESS); |
| 143 | udelay(30); /* It takes 30 us to read/write the data */ |
| 144 | } while (--repeat && (reg_val & PHY_CMD_ACTIVE)); |
| 145 | |
| 146 | if(reg_val & PHY_RD_ERR) |
| 147 | goto err_phy_write; |
| 148 | |
| 149 | return 0; |
| 150 | |
| 151 | err_phy_write: |
| 152 | return -EINVAL; |
| 153 | |
| 154 | } |
| 155 | |
| 156 | /* This is the mii register read function provided to the mii interface. */ |
| 157 | static int amd8111e_mdio_read(struct net_device *dev, int phy_id, int reg_num) |
| 158 | { |
| 159 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 160 | unsigned int reg_val; |
| 161 | |
| 162 | amd8111e_read_phy(lp,phy_id,reg_num,®_val); |
| 163 | return reg_val; |
| 164 | |
| 165 | } |
| 166 | |
| 167 | /* This is the mii register write function provided to the mii interface. */ |
| 168 | static void amd8111e_mdio_write(struct net_device *dev, |
| 169 | int phy_id, int reg_num, int val) |
| 170 | { |
| 171 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 172 | |
| 173 | amd8111e_write_phy(lp, phy_id, reg_num, val); |
| 174 | } |
| 175 | |
| 176 | /* This function will set PHY speed. During initialization sets |
| 177 | * the original speed to 100 full |
| 178 | */ |
| 179 | static void amd8111e_set_ext_phy(struct net_device *dev) |
| 180 | { |
| 181 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 182 | u32 bmcr,advert,tmp; |
| 183 | |
| 184 | /* Determine mii register values to set the speed */ |
| 185 | advert = amd8111e_mdio_read(dev, lp->ext_phy_addr, MII_ADVERTISE); |
| 186 | tmp = advert & ~(ADVERTISE_ALL | ADVERTISE_100BASE4); |
| 187 | switch (lp->ext_phy_option){ |
| 188 | |
| 189 | default: |
| 190 | case SPEED_AUTONEG: /* advertise all values */ |
| 191 | tmp |= ( ADVERTISE_10HALF|ADVERTISE_10FULL| |
| 192 | ADVERTISE_100HALF|ADVERTISE_100FULL) ; |
| 193 | break; |
| 194 | case SPEED10_HALF: |
| 195 | tmp |= ADVERTISE_10HALF; |
| 196 | break; |
| 197 | case SPEED10_FULL: |
| 198 | tmp |= ADVERTISE_10FULL; |
| 199 | break; |
| 200 | case SPEED100_HALF: |
| 201 | tmp |= ADVERTISE_100HALF; |
| 202 | break; |
| 203 | case SPEED100_FULL: |
| 204 | tmp |= ADVERTISE_100FULL; |
| 205 | break; |
| 206 | } |
| 207 | |
| 208 | if(advert != tmp) |
| 209 | amd8111e_mdio_write(dev, lp->ext_phy_addr, MII_ADVERTISE, tmp); |
| 210 | /* Restart auto negotiation */ |
| 211 | bmcr = amd8111e_mdio_read(dev, lp->ext_phy_addr, MII_BMCR); |
| 212 | bmcr |= (BMCR_ANENABLE | BMCR_ANRESTART); |
| 213 | amd8111e_mdio_write(dev, lp->ext_phy_addr, MII_BMCR, bmcr); |
| 214 | |
| 215 | } |
| 216 | |
| 217 | /* This function will unmap skb->data space and will free |
| 218 | * all transmit and receive skbuffs. |
| 219 | */ |
| 220 | static int amd8111e_free_skbs(struct net_device *dev) |
| 221 | { |
| 222 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 223 | struct sk_buff *rx_skbuff; |
| 224 | int i; |
| 225 | |
| 226 | /* Freeing transmit skbs */ |
| 227 | for(i = 0; i < NUM_TX_BUFFERS; i++){ |
| 228 | if(lp->tx_skbuff[i]){ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 229 | dma_unmap_single(&lp->pci_dev->dev, |
| 230 | lp->tx_dma_addr[i], |
| 231 | lp->tx_skbuff[i]->len, DMA_TO_DEVICE); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 232 | dev_kfree_skb (lp->tx_skbuff[i]); |
| 233 | lp->tx_skbuff[i] = NULL; |
| 234 | lp->tx_dma_addr[i] = 0; |
| 235 | } |
| 236 | } |
| 237 | /* Freeing previously allocated receive buffers */ |
| 238 | for (i = 0; i < NUM_RX_BUFFERS; i++){ |
| 239 | rx_skbuff = lp->rx_skbuff[i]; |
| 240 | if(rx_skbuff != NULL){ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 241 | dma_unmap_single(&lp->pci_dev->dev, |
| 242 | lp->rx_dma_addr[i], |
| 243 | lp->rx_buff_len - 2, DMA_FROM_DEVICE); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 244 | dev_kfree_skb(lp->rx_skbuff[i]); |
| 245 | lp->rx_skbuff[i] = NULL; |
| 246 | lp->rx_dma_addr[i] = 0; |
| 247 | } |
| 248 | } |
| 249 | |
| 250 | return 0; |
| 251 | } |
| 252 | |
| 253 | /* This will set the receive buffer length corresponding |
| 254 | * to the mtu size of networkinterface. |
| 255 | */ |
| 256 | static inline void amd8111e_set_rx_buff_len(struct net_device *dev) |
| 257 | { |
| 258 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 259 | unsigned int mtu = dev->mtu; |
| 260 | |
| 261 | if (mtu > ETH_DATA_LEN){ |
| 262 | /* MTU + ethernet header + FCS |
| 263 | * + optional VLAN tag + skb reserve space 2 |
| 264 | */ |
| 265 | lp->rx_buff_len = mtu + ETH_HLEN + 10; |
| 266 | lp->options |= OPTION_JUMBO_ENABLE; |
| 267 | } else{ |
| 268 | lp->rx_buff_len = PKT_BUFF_SZ; |
| 269 | lp->options &= ~OPTION_JUMBO_ENABLE; |
| 270 | } |
| 271 | } |
| 272 | |
| 273 | /* This function will free all the previously allocated buffers, |
| 274 | * determine new receive buffer length and will allocate new receive buffers. |
| 275 | * This function also allocates and initializes both the transmitter |
| 276 | * and receive hardware descriptors. |
| 277 | */ |
| 278 | static int amd8111e_init_ring(struct net_device *dev) |
| 279 | { |
| 280 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 281 | int i; |
| 282 | |
| 283 | lp->rx_idx = lp->tx_idx = 0; |
| 284 | lp->tx_complete_idx = 0; |
| 285 | lp->tx_ring_idx = 0; |
| 286 | |
| 287 | |
| 288 | if(lp->opened) |
| 289 | /* Free previously allocated transmit and receive skbs */ |
| 290 | amd8111e_free_skbs(dev); |
| 291 | |
| 292 | else{ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 293 | /* allocate the tx and rx descriptors */ |
| 294 | lp->tx_ring = dma_alloc_coherent(&lp->pci_dev->dev, |
| 295 | sizeof(struct amd8111e_tx_dr) * NUM_TX_RING_DR, |
| 296 | &lp->tx_ring_dma_addr, GFP_ATOMIC); |
| 297 | if (!lp->tx_ring) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 298 | goto err_no_mem; |
| 299 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 300 | lp->rx_ring = dma_alloc_coherent(&lp->pci_dev->dev, |
| 301 | sizeof(struct amd8111e_rx_dr) * NUM_RX_RING_DR, |
| 302 | &lp->rx_ring_dma_addr, GFP_ATOMIC); |
| 303 | if (!lp->rx_ring) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 304 | goto err_free_tx_ring; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 305 | } |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 306 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 307 | /* Set new receive buff size */ |
| 308 | amd8111e_set_rx_buff_len(dev); |
| 309 | |
| 310 | /* Allocating receive skbs */ |
| 311 | for (i = 0; i < NUM_RX_BUFFERS; i++) { |
| 312 | |
| 313 | lp->rx_skbuff[i] = netdev_alloc_skb(dev, lp->rx_buff_len); |
| 314 | if (!lp->rx_skbuff[i]) { |
| 315 | /* Release previos allocated skbs */ |
| 316 | for(--i; i >= 0 ;i--) |
| 317 | dev_kfree_skb(lp->rx_skbuff[i]); |
| 318 | goto err_free_rx_ring; |
| 319 | } |
| 320 | skb_reserve(lp->rx_skbuff[i],2); |
| 321 | } |
| 322 | /* Initilaizing receive descriptors */ |
| 323 | for (i = 0; i < NUM_RX_BUFFERS; i++) { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 324 | lp->rx_dma_addr[i] = dma_map_single(&lp->pci_dev->dev, |
| 325 | lp->rx_skbuff[i]->data, |
| 326 | lp->rx_buff_len - 2, |
| 327 | DMA_FROM_DEVICE); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 328 | |
| 329 | lp->rx_ring[i].buff_phy_addr = cpu_to_le32(lp->rx_dma_addr[i]); |
| 330 | lp->rx_ring[i].buff_count = cpu_to_le16(lp->rx_buff_len-2); |
| 331 | wmb(); |
| 332 | lp->rx_ring[i].rx_flags = cpu_to_le16(OWN_BIT); |
| 333 | } |
| 334 | |
| 335 | /* Initializing transmit descriptors */ |
| 336 | for (i = 0; i < NUM_TX_RING_DR; i++) { |
| 337 | lp->tx_ring[i].buff_phy_addr = 0; |
| 338 | lp->tx_ring[i].tx_flags = 0; |
| 339 | lp->tx_ring[i].buff_count = 0; |
| 340 | } |
| 341 | |
| 342 | return 0; |
| 343 | |
| 344 | err_free_rx_ring: |
| 345 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 346 | dma_free_coherent(&lp->pci_dev->dev, |
| 347 | sizeof(struct amd8111e_rx_dr) * NUM_RX_RING_DR, |
| 348 | lp->rx_ring, lp->rx_ring_dma_addr); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 349 | |
| 350 | err_free_tx_ring: |
| 351 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 352 | dma_free_coherent(&lp->pci_dev->dev, |
| 353 | sizeof(struct amd8111e_tx_dr) * NUM_TX_RING_DR, |
| 354 | lp->tx_ring, lp->tx_ring_dma_addr); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 355 | |
| 356 | err_no_mem: |
| 357 | return -ENOMEM; |
| 358 | } |
| 359 | |
| 360 | /* This function will set the interrupt coalescing according |
| 361 | * to the input arguments |
| 362 | */ |
| 363 | static int amd8111e_set_coalesce(struct net_device *dev, enum coal_mode cmod) |
| 364 | { |
| 365 | unsigned int timeout; |
| 366 | unsigned int event_count; |
| 367 | |
| 368 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 369 | void __iomem *mmio = lp->mmio; |
| 370 | struct amd8111e_coalesce_conf *coal_conf = &lp->coal_conf; |
| 371 | |
| 372 | |
| 373 | switch(cmod) |
| 374 | { |
| 375 | case RX_INTR_COAL : |
| 376 | timeout = coal_conf->rx_timeout; |
| 377 | event_count = coal_conf->rx_event_count; |
| 378 | if( timeout > MAX_TIMEOUT || |
| 379 | event_count > MAX_EVENT_COUNT ) |
| 380 | return -EINVAL; |
| 381 | |
| 382 | timeout = timeout * DELAY_TIMER_CONV; |
| 383 | writel(VAL0|STINTEN, mmio+INTEN0); |
| 384 | writel((u32)DLY_INT_A_R0|( event_count<< 16 )|timeout, |
| 385 | mmio+DLY_INT_A); |
| 386 | break; |
| 387 | |
| 388 | case TX_INTR_COAL : |
| 389 | timeout = coal_conf->tx_timeout; |
| 390 | event_count = coal_conf->tx_event_count; |
| 391 | if( timeout > MAX_TIMEOUT || |
| 392 | event_count > MAX_EVENT_COUNT ) |
| 393 | return -EINVAL; |
| 394 | |
| 395 | |
| 396 | timeout = timeout * DELAY_TIMER_CONV; |
| 397 | writel(VAL0|STINTEN,mmio+INTEN0); |
| 398 | writel((u32)DLY_INT_B_T0|( event_count<< 16 )|timeout, |
| 399 | mmio+DLY_INT_B); |
| 400 | break; |
| 401 | |
| 402 | case DISABLE_COAL: |
| 403 | writel(0,mmio+STVAL); |
| 404 | writel(STINTEN, mmio+INTEN0); |
| 405 | writel(0, mmio +DLY_INT_B); |
| 406 | writel(0, mmio+DLY_INT_A); |
| 407 | break; |
| 408 | case ENABLE_COAL: |
| 409 | /* Start the timer */ |
| 410 | writel((u32)SOFT_TIMER_FREQ, mmio+STVAL); /* 0.5 sec */ |
| 411 | writel(VAL0|STINTEN, mmio+INTEN0); |
| 412 | break; |
| 413 | default: |
| 414 | break; |
| 415 | |
| 416 | } |
| 417 | return 0; |
| 418 | |
| 419 | } |
| 420 | |
| 421 | /* This function initializes the device registers and starts the device. */ |
| 422 | static int amd8111e_restart(struct net_device *dev) |
| 423 | { |
| 424 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 425 | void __iomem *mmio = lp->mmio; |
| 426 | int i,reg_val; |
| 427 | |
| 428 | /* stop the chip */ |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 429 | writel(RUN, mmio + CMD0); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 430 | |
| 431 | if(amd8111e_init_ring(dev)) |
| 432 | return -ENOMEM; |
| 433 | |
| 434 | /* enable the port manager and set auto negotiation always */ |
| 435 | writel((u32) VAL1|EN_PMGR, mmio + CMD3 ); |
| 436 | writel((u32)XPHYANE|XPHYRST , mmio + CTRL2); |
| 437 | |
| 438 | amd8111e_set_ext_phy(dev); |
| 439 | |
| 440 | /* set control registers */ |
| 441 | reg_val = readl(mmio + CTRL1); |
| 442 | reg_val &= ~XMTSP_MASK; |
| 443 | writel( reg_val| XMTSP_128 | CACHE_ALIGN, mmio + CTRL1 ); |
| 444 | |
| 445 | /* enable interrupt */ |
| 446 | writel( APINT5EN | APINT4EN | APINT3EN | APINT2EN | APINT1EN | |
| 447 | APINT0EN | MIIPDTINTEN | MCCIINTEN | MCCINTEN | MREINTEN | |
| 448 | SPNDINTEN | MPINTEN | SINTEN | STINTEN, mmio + INTEN0); |
| 449 | |
| 450 | writel(VAL3 | LCINTEN | VAL1 | TINTEN0 | VAL0 | RINTEN0, mmio + INTEN0); |
| 451 | |
| 452 | /* initialize tx and rx ring base addresses */ |
| 453 | writel((u32)lp->tx_ring_dma_addr,mmio + XMT_RING_BASE_ADDR0); |
| 454 | writel((u32)lp->rx_ring_dma_addr,mmio+ RCV_RING_BASE_ADDR0); |
| 455 | |
| 456 | writew((u32)NUM_TX_RING_DR, mmio + XMT_RING_LEN0); |
| 457 | writew((u16)NUM_RX_RING_DR, mmio + RCV_RING_LEN0); |
| 458 | |
| 459 | /* set default IPG to 96 */ |
| 460 | writew((u32)DEFAULT_IPG,mmio+IPG); |
| 461 | writew((u32)(DEFAULT_IPG-IFS1_DELTA), mmio + IFS1); |
| 462 | |
| 463 | if(lp->options & OPTION_JUMBO_ENABLE){ |
| 464 | writel((u32)VAL2|JUMBO, mmio + CMD3); |
| 465 | /* Reset REX_UFLO */ |
| 466 | writel( REX_UFLO, mmio + CMD2); |
| 467 | /* Should not set REX_UFLO for jumbo frames */ |
| 468 | writel( VAL0 | APAD_XMT|REX_RTRY , mmio + CMD2); |
| 469 | }else{ |
| 470 | writel( VAL0 | APAD_XMT | REX_RTRY|REX_UFLO, mmio + CMD2); |
| 471 | writel((u32)JUMBO, mmio + CMD3); |
| 472 | } |
| 473 | |
| 474 | #if AMD8111E_VLAN_TAG_USED |
| 475 | writel((u32) VAL2|VSIZE|VL_TAG_DEL, mmio + CMD3); |
| 476 | #endif |
| 477 | writel( VAL0 | APAD_XMT | REX_RTRY, mmio + CMD2 ); |
| 478 | |
| 479 | /* Setting the MAC address to the device */ |
| 480 | for (i = 0; i < ETH_ALEN; i++) |
| 481 | writeb( dev->dev_addr[i], mmio + PADR + i ); |
| 482 | |
| 483 | /* Enable interrupt coalesce */ |
| 484 | if(lp->options & OPTION_INTR_COAL_ENABLE){ |
| 485 | netdev_info(dev, "Interrupt Coalescing Enabled.\n"); |
| 486 | amd8111e_set_coalesce(dev,ENABLE_COAL); |
| 487 | } |
| 488 | |
| 489 | /* set RUN bit to start the chip */ |
| 490 | writel(VAL2 | RDMD0, mmio + CMD0); |
| 491 | writel(VAL0 | INTREN | RUN, mmio + CMD0); |
| 492 | |
| 493 | /* To avoid PCI posting bug */ |
| 494 | readl(mmio+CMD0); |
| 495 | return 0; |
| 496 | } |
| 497 | |
| 498 | /* This function clears necessary the device registers. */ |
| 499 | static void amd8111e_init_hw_default(struct amd8111e_priv *lp) |
| 500 | { |
| 501 | unsigned int reg_val; |
| 502 | unsigned int logic_filter[2] ={0,}; |
| 503 | void __iomem *mmio = lp->mmio; |
| 504 | |
| 505 | |
| 506 | /* stop the chip */ |
| 507 | writel(RUN, mmio + CMD0); |
| 508 | |
| 509 | /* AUTOPOLL0 Register *//*TBD default value is 8100 in FPS */ |
| 510 | writew( 0x8100 | lp->ext_phy_addr, mmio + AUTOPOLL0); |
| 511 | |
| 512 | /* Clear RCV_RING_BASE_ADDR */ |
| 513 | writel(0, mmio + RCV_RING_BASE_ADDR0); |
| 514 | |
| 515 | /* Clear XMT_RING_BASE_ADDR */ |
| 516 | writel(0, mmio + XMT_RING_BASE_ADDR0); |
| 517 | writel(0, mmio + XMT_RING_BASE_ADDR1); |
| 518 | writel(0, mmio + XMT_RING_BASE_ADDR2); |
| 519 | writel(0, mmio + XMT_RING_BASE_ADDR3); |
| 520 | |
| 521 | /* Clear CMD0 */ |
| 522 | writel(CMD0_CLEAR,mmio + CMD0); |
| 523 | |
| 524 | /* Clear CMD2 */ |
| 525 | writel(CMD2_CLEAR, mmio +CMD2); |
| 526 | |
| 527 | /* Clear CMD7 */ |
| 528 | writel(CMD7_CLEAR , mmio + CMD7); |
| 529 | |
| 530 | /* Clear DLY_INT_A and DLY_INT_B */ |
| 531 | writel(0x0, mmio + DLY_INT_A); |
| 532 | writel(0x0, mmio + DLY_INT_B); |
| 533 | |
| 534 | /* Clear FLOW_CONTROL */ |
| 535 | writel(0x0, mmio + FLOW_CONTROL); |
| 536 | |
| 537 | /* Clear INT0 write 1 to clear register */ |
| 538 | reg_val = readl(mmio + INT0); |
| 539 | writel(reg_val, mmio + INT0); |
| 540 | |
| 541 | /* Clear STVAL */ |
| 542 | writel(0x0, mmio + STVAL); |
| 543 | |
| 544 | /* Clear INTEN0 */ |
| 545 | writel( INTEN0_CLEAR, mmio + INTEN0); |
| 546 | |
| 547 | /* Clear LADRF */ |
| 548 | writel(0x0 , mmio + LADRF); |
| 549 | |
| 550 | /* Set SRAM_SIZE & SRAM_BOUNDARY registers */ |
| 551 | writel( 0x80010,mmio + SRAM_SIZE); |
| 552 | |
| 553 | /* Clear RCV_RING0_LEN */ |
| 554 | writel(0x0, mmio + RCV_RING_LEN0); |
| 555 | |
| 556 | /* Clear XMT_RING0/1/2/3_LEN */ |
| 557 | writel(0x0, mmio + XMT_RING_LEN0); |
| 558 | writel(0x0, mmio + XMT_RING_LEN1); |
| 559 | writel(0x0, mmio + XMT_RING_LEN2); |
| 560 | writel(0x0, mmio + XMT_RING_LEN3); |
| 561 | |
| 562 | /* Clear XMT_RING_LIMIT */ |
| 563 | writel(0x0, mmio + XMT_RING_LIMIT); |
| 564 | |
| 565 | /* Clear MIB */ |
| 566 | writew(MIB_CLEAR, mmio + MIB_ADDR); |
| 567 | |
| 568 | /* Clear LARF */ |
| 569 | amd8111e_writeq(*(u64 *)logic_filter, mmio + LADRF); |
| 570 | |
| 571 | /* SRAM_SIZE register */ |
| 572 | reg_val = readl(mmio + SRAM_SIZE); |
| 573 | |
| 574 | if(lp->options & OPTION_JUMBO_ENABLE) |
| 575 | writel( VAL2|JUMBO, mmio + CMD3); |
| 576 | #if AMD8111E_VLAN_TAG_USED |
| 577 | writel(VAL2|VSIZE|VL_TAG_DEL, mmio + CMD3 ); |
| 578 | #endif |
| 579 | /* Set default value to CTRL1 Register */ |
| 580 | writel(CTRL1_DEFAULT, mmio + CTRL1); |
| 581 | |
| 582 | /* To avoid PCI posting bug */ |
| 583 | readl(mmio + CMD2); |
| 584 | |
| 585 | } |
| 586 | |
| 587 | /* This function disables the interrupt and clears all the pending |
| 588 | * interrupts in INT0 |
| 589 | */ |
| 590 | static void amd8111e_disable_interrupt(struct amd8111e_priv *lp) |
| 591 | { |
| 592 | u32 intr0; |
| 593 | |
| 594 | /* Disable interrupt */ |
| 595 | writel(INTREN, lp->mmio + CMD0); |
| 596 | |
| 597 | /* Clear INT0 */ |
| 598 | intr0 = readl(lp->mmio + INT0); |
| 599 | writel(intr0, lp->mmio + INT0); |
| 600 | |
| 601 | /* To avoid PCI posting bug */ |
| 602 | readl(lp->mmio + INT0); |
| 603 | |
| 604 | } |
| 605 | |
| 606 | /* This function stops the chip. */ |
| 607 | static void amd8111e_stop_chip(struct amd8111e_priv *lp) |
| 608 | { |
| 609 | writel(RUN, lp->mmio + CMD0); |
| 610 | |
| 611 | /* To avoid PCI posting bug */ |
| 612 | readl(lp->mmio + CMD0); |
| 613 | } |
| 614 | |
| 615 | /* This function frees the transmiter and receiver descriptor rings. */ |
| 616 | static void amd8111e_free_ring(struct amd8111e_priv *lp) |
| 617 | { |
| 618 | /* Free transmit and receive descriptor rings */ |
| 619 | if(lp->rx_ring){ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 620 | dma_free_coherent(&lp->pci_dev->dev, |
| 621 | sizeof(struct amd8111e_rx_dr) * NUM_RX_RING_DR, |
| 622 | lp->rx_ring, lp->rx_ring_dma_addr); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 623 | lp->rx_ring = NULL; |
| 624 | } |
| 625 | |
| 626 | if(lp->tx_ring){ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 627 | dma_free_coherent(&lp->pci_dev->dev, |
| 628 | sizeof(struct amd8111e_tx_dr) * NUM_TX_RING_DR, |
| 629 | lp->tx_ring, lp->tx_ring_dma_addr); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 630 | |
| 631 | lp->tx_ring = NULL; |
| 632 | } |
| 633 | |
| 634 | } |
| 635 | |
| 636 | /* This function will free all the transmit skbs that are actually |
| 637 | * transmitted by the device. It will check the ownership of the |
| 638 | * skb before freeing the skb. |
| 639 | */ |
| 640 | static int amd8111e_tx(struct net_device *dev) |
| 641 | { |
| 642 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 643 | int tx_index; |
| 644 | int status; |
| 645 | /* Complete all the transmit packet */ |
| 646 | while (lp->tx_complete_idx != lp->tx_idx){ |
| 647 | tx_index = lp->tx_complete_idx & TX_RING_DR_MOD_MASK; |
| 648 | status = le16_to_cpu(lp->tx_ring[tx_index].tx_flags); |
| 649 | |
| 650 | if(status & OWN_BIT) |
| 651 | break; /* It still hasn't been Txed */ |
| 652 | |
| 653 | lp->tx_ring[tx_index].buff_phy_addr = 0; |
| 654 | |
| 655 | /* We must free the original skb */ |
| 656 | if (lp->tx_skbuff[tx_index]) { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 657 | dma_unmap_single(&lp->pci_dev->dev, |
| 658 | lp->tx_dma_addr[tx_index], |
| 659 | lp->tx_skbuff[tx_index]->len, |
| 660 | DMA_TO_DEVICE); |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 661 | dev_consume_skb_irq(lp->tx_skbuff[tx_index]); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 662 | lp->tx_skbuff[tx_index] = NULL; |
| 663 | lp->tx_dma_addr[tx_index] = 0; |
| 664 | } |
| 665 | lp->tx_complete_idx++; |
| 666 | /*COAL update tx coalescing parameters */ |
| 667 | lp->coal_conf.tx_packets++; |
| 668 | lp->coal_conf.tx_bytes += |
| 669 | le16_to_cpu(lp->tx_ring[tx_index].buff_count); |
| 670 | |
| 671 | if (netif_queue_stopped(dev) && |
| 672 | lp->tx_complete_idx > lp->tx_idx - NUM_TX_BUFFERS +2){ |
| 673 | /* The ring is no longer full, clear tbusy. */ |
| 674 | /* lp->tx_full = 0; */ |
| 675 | netif_wake_queue (dev); |
| 676 | } |
| 677 | } |
| 678 | return 0; |
| 679 | } |
| 680 | |
| 681 | /* This function handles the driver receive operation in polling mode */ |
| 682 | static int amd8111e_rx_poll(struct napi_struct *napi, int budget) |
| 683 | { |
| 684 | struct amd8111e_priv *lp = container_of(napi, struct amd8111e_priv, napi); |
| 685 | struct net_device *dev = lp->amd8111e_net_dev; |
| 686 | int rx_index = lp->rx_idx & RX_RING_DR_MOD_MASK; |
| 687 | void __iomem *mmio = lp->mmio; |
| 688 | struct sk_buff *skb,*new_skb; |
| 689 | int min_pkt_len, status; |
| 690 | int num_rx_pkt = 0; |
| 691 | short pkt_len; |
| 692 | #if AMD8111E_VLAN_TAG_USED |
| 693 | short vtag; |
| 694 | #endif |
| 695 | |
| 696 | while (num_rx_pkt < budget) { |
| 697 | status = le16_to_cpu(lp->rx_ring[rx_index].rx_flags); |
| 698 | if (status & OWN_BIT) |
| 699 | break; |
| 700 | |
| 701 | /* There is a tricky error noted by John Murphy, |
| 702 | * <murf@perftech.com> to Russ Nelson: Even with |
| 703 | * full-sized * buffers it's possible for a |
| 704 | * jabber packet to use two buffers, with only |
| 705 | * the last correctly noting the error. |
| 706 | */ |
| 707 | if (status & ERR_BIT) { |
| 708 | /* resetting flags */ |
| 709 | lp->rx_ring[rx_index].rx_flags &= RESET_RX_FLAGS; |
| 710 | goto err_next_pkt; |
| 711 | } |
| 712 | /* check for STP and ENP */ |
| 713 | if (!((status & STP_BIT) && (status & ENP_BIT))){ |
| 714 | /* resetting flags */ |
| 715 | lp->rx_ring[rx_index].rx_flags &= RESET_RX_FLAGS; |
| 716 | goto err_next_pkt; |
| 717 | } |
| 718 | pkt_len = le16_to_cpu(lp->rx_ring[rx_index].msg_count) - 4; |
| 719 | |
| 720 | #if AMD8111E_VLAN_TAG_USED |
| 721 | vtag = status & TT_MASK; |
| 722 | /* MAC will strip vlan tag */ |
| 723 | if (vtag != 0) |
| 724 | min_pkt_len = MIN_PKT_LEN - 4; |
| 725 | else |
| 726 | #endif |
| 727 | min_pkt_len = MIN_PKT_LEN; |
| 728 | |
| 729 | if (pkt_len < min_pkt_len) { |
| 730 | lp->rx_ring[rx_index].rx_flags &= RESET_RX_FLAGS; |
| 731 | lp->drv_rx_errors++; |
| 732 | goto err_next_pkt; |
| 733 | } |
| 734 | new_skb = netdev_alloc_skb(dev, lp->rx_buff_len); |
| 735 | if (!new_skb) { |
| 736 | /* if allocation fail, |
| 737 | * ignore that pkt and go to next one |
| 738 | */ |
| 739 | lp->rx_ring[rx_index].rx_flags &= RESET_RX_FLAGS; |
| 740 | lp->drv_rx_errors++; |
| 741 | goto err_next_pkt; |
| 742 | } |
| 743 | |
| 744 | skb_reserve(new_skb, 2); |
| 745 | skb = lp->rx_skbuff[rx_index]; |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 746 | dma_unmap_single(&lp->pci_dev->dev, lp->rx_dma_addr[rx_index], |
| 747 | lp->rx_buff_len - 2, DMA_FROM_DEVICE); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 748 | skb_put(skb, pkt_len); |
| 749 | lp->rx_skbuff[rx_index] = new_skb; |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 750 | lp->rx_dma_addr[rx_index] = dma_map_single(&lp->pci_dev->dev, |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 751 | new_skb->data, |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 752 | lp->rx_buff_len - 2, |
| 753 | DMA_FROM_DEVICE); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 754 | |
| 755 | skb->protocol = eth_type_trans(skb, dev); |
| 756 | |
| 757 | #if AMD8111E_VLAN_TAG_USED |
| 758 | if (vtag == TT_VLAN_TAGGED){ |
| 759 | u16 vlan_tag = le16_to_cpu(lp->rx_ring[rx_index].tag_ctrl_info); |
| 760 | __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), vlan_tag); |
| 761 | } |
| 762 | #endif |
| 763 | napi_gro_receive(napi, skb); |
| 764 | /* COAL update rx coalescing parameters */ |
| 765 | lp->coal_conf.rx_packets++; |
| 766 | lp->coal_conf.rx_bytes += pkt_len; |
| 767 | num_rx_pkt++; |
| 768 | |
| 769 | err_next_pkt: |
| 770 | lp->rx_ring[rx_index].buff_phy_addr |
| 771 | = cpu_to_le32(lp->rx_dma_addr[rx_index]); |
| 772 | lp->rx_ring[rx_index].buff_count = |
| 773 | cpu_to_le16(lp->rx_buff_len-2); |
| 774 | wmb(); |
| 775 | lp->rx_ring[rx_index].rx_flags |= cpu_to_le16(OWN_BIT); |
| 776 | rx_index = (++lp->rx_idx) & RX_RING_DR_MOD_MASK; |
| 777 | } |
| 778 | |
| 779 | if (num_rx_pkt < budget && napi_complete_done(napi, num_rx_pkt)) { |
| 780 | unsigned long flags; |
| 781 | |
| 782 | /* Receive descriptor is empty now */ |
| 783 | spin_lock_irqsave(&lp->lock, flags); |
| 784 | writel(VAL0|RINTEN0, mmio + INTEN0); |
| 785 | writel(VAL2 | RDMD0, mmio + CMD0); |
| 786 | spin_unlock_irqrestore(&lp->lock, flags); |
| 787 | } |
| 788 | |
| 789 | return num_rx_pkt; |
| 790 | } |
| 791 | |
| 792 | /* This function will indicate the link status to the kernel. */ |
| 793 | static int amd8111e_link_change(struct net_device *dev) |
| 794 | { |
| 795 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 796 | int status0,speed; |
| 797 | |
| 798 | /* read the link change */ |
| 799 | status0 = readl(lp->mmio + STAT0); |
| 800 | |
| 801 | if(status0 & LINK_STATS){ |
| 802 | if(status0 & AUTONEG_COMPLETE) |
| 803 | lp->link_config.autoneg = AUTONEG_ENABLE; |
| 804 | else |
| 805 | lp->link_config.autoneg = AUTONEG_DISABLE; |
| 806 | |
| 807 | if(status0 & FULL_DPLX) |
| 808 | lp->link_config.duplex = DUPLEX_FULL; |
| 809 | else |
| 810 | lp->link_config.duplex = DUPLEX_HALF; |
| 811 | speed = (status0 & SPEED_MASK) >> 7; |
| 812 | if(speed == PHY_SPEED_10) |
| 813 | lp->link_config.speed = SPEED_10; |
| 814 | else if(speed == PHY_SPEED_100) |
| 815 | lp->link_config.speed = SPEED_100; |
| 816 | |
| 817 | netdev_info(dev, "Link is Up. Speed is %s Mbps %s Duplex\n", |
| 818 | (lp->link_config.speed == SPEED_100) ? |
| 819 | "100" : "10", |
| 820 | (lp->link_config.duplex == DUPLEX_FULL) ? |
| 821 | "Full" : "Half"); |
| 822 | |
| 823 | netif_carrier_on(dev); |
| 824 | } |
| 825 | else{ |
| 826 | lp->link_config.speed = SPEED_INVALID; |
| 827 | lp->link_config.duplex = DUPLEX_INVALID; |
| 828 | lp->link_config.autoneg = AUTONEG_INVALID; |
| 829 | netdev_info(dev, "Link is Down.\n"); |
| 830 | netif_carrier_off(dev); |
| 831 | } |
| 832 | |
| 833 | return 0; |
| 834 | } |
| 835 | |
| 836 | /* This function reads the mib counters. */ |
| 837 | static int amd8111e_read_mib(void __iomem *mmio, u8 MIB_COUNTER) |
| 838 | { |
| 839 | unsigned int status; |
| 840 | unsigned int data; |
| 841 | unsigned int repeat = REPEAT_CNT; |
| 842 | |
| 843 | writew( MIB_RD_CMD | MIB_COUNTER, mmio + MIB_ADDR); |
| 844 | do { |
| 845 | status = readw(mmio + MIB_ADDR); |
| 846 | udelay(2); /* controller takes MAX 2 us to get mib data */ |
| 847 | } |
| 848 | while (--repeat && (status & MIB_CMD_ACTIVE)); |
| 849 | |
| 850 | data = readl(mmio + MIB_DATA); |
| 851 | return data; |
| 852 | } |
| 853 | |
| 854 | /* This function reads the mib registers and returns the hardware statistics. |
| 855 | * It updates previous internal driver statistics with new values. |
| 856 | */ |
| 857 | static struct net_device_stats *amd8111e_get_stats(struct net_device *dev) |
| 858 | { |
| 859 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 860 | void __iomem *mmio = lp->mmio; |
| 861 | unsigned long flags; |
| 862 | struct net_device_stats *new_stats = &dev->stats; |
| 863 | |
| 864 | if (!lp->opened) |
| 865 | return new_stats; |
| 866 | spin_lock_irqsave (&lp->lock, flags); |
| 867 | |
| 868 | /* stats.rx_packets */ |
| 869 | new_stats->rx_packets = amd8111e_read_mib(mmio, rcv_broadcast_pkts)+ |
| 870 | amd8111e_read_mib(mmio, rcv_multicast_pkts)+ |
| 871 | amd8111e_read_mib(mmio, rcv_unicast_pkts); |
| 872 | |
| 873 | /* stats.tx_packets */ |
| 874 | new_stats->tx_packets = amd8111e_read_mib(mmio, xmt_packets); |
| 875 | |
| 876 | /*stats.rx_bytes */ |
| 877 | new_stats->rx_bytes = amd8111e_read_mib(mmio, rcv_octets); |
| 878 | |
| 879 | /* stats.tx_bytes */ |
| 880 | new_stats->tx_bytes = amd8111e_read_mib(mmio, xmt_octets); |
| 881 | |
| 882 | /* stats.rx_errors */ |
| 883 | /* hw errors + errors driver reported */ |
| 884 | new_stats->rx_errors = amd8111e_read_mib(mmio, rcv_undersize_pkts)+ |
| 885 | amd8111e_read_mib(mmio, rcv_fragments)+ |
| 886 | amd8111e_read_mib(mmio, rcv_jabbers)+ |
| 887 | amd8111e_read_mib(mmio, rcv_alignment_errors)+ |
| 888 | amd8111e_read_mib(mmio, rcv_fcs_errors)+ |
| 889 | amd8111e_read_mib(mmio, rcv_miss_pkts)+ |
| 890 | lp->drv_rx_errors; |
| 891 | |
| 892 | /* stats.tx_errors */ |
| 893 | new_stats->tx_errors = amd8111e_read_mib(mmio, xmt_underrun_pkts); |
| 894 | |
| 895 | /* stats.rx_dropped*/ |
| 896 | new_stats->rx_dropped = amd8111e_read_mib(mmio, rcv_miss_pkts); |
| 897 | |
| 898 | /* stats.tx_dropped*/ |
| 899 | new_stats->tx_dropped = amd8111e_read_mib(mmio, xmt_underrun_pkts); |
| 900 | |
| 901 | /* stats.multicast*/ |
| 902 | new_stats->multicast = amd8111e_read_mib(mmio, rcv_multicast_pkts); |
| 903 | |
| 904 | /* stats.collisions*/ |
| 905 | new_stats->collisions = amd8111e_read_mib(mmio, xmt_collisions); |
| 906 | |
| 907 | /* stats.rx_length_errors*/ |
| 908 | new_stats->rx_length_errors = |
| 909 | amd8111e_read_mib(mmio, rcv_undersize_pkts)+ |
| 910 | amd8111e_read_mib(mmio, rcv_oversize_pkts); |
| 911 | |
| 912 | /* stats.rx_over_errors*/ |
| 913 | new_stats->rx_over_errors = amd8111e_read_mib(mmio, rcv_miss_pkts); |
| 914 | |
| 915 | /* stats.rx_crc_errors*/ |
| 916 | new_stats->rx_crc_errors = amd8111e_read_mib(mmio, rcv_fcs_errors); |
| 917 | |
| 918 | /* stats.rx_frame_errors*/ |
| 919 | new_stats->rx_frame_errors = |
| 920 | amd8111e_read_mib(mmio, rcv_alignment_errors); |
| 921 | |
| 922 | /* stats.rx_fifo_errors */ |
| 923 | new_stats->rx_fifo_errors = amd8111e_read_mib(mmio, rcv_miss_pkts); |
| 924 | |
| 925 | /* stats.rx_missed_errors */ |
| 926 | new_stats->rx_missed_errors = amd8111e_read_mib(mmio, rcv_miss_pkts); |
| 927 | |
| 928 | /* stats.tx_aborted_errors*/ |
| 929 | new_stats->tx_aborted_errors = |
| 930 | amd8111e_read_mib(mmio, xmt_excessive_collision); |
| 931 | |
| 932 | /* stats.tx_carrier_errors*/ |
| 933 | new_stats->tx_carrier_errors = |
| 934 | amd8111e_read_mib(mmio, xmt_loss_carrier); |
| 935 | |
| 936 | /* stats.tx_fifo_errors*/ |
| 937 | new_stats->tx_fifo_errors = amd8111e_read_mib(mmio, xmt_underrun_pkts); |
| 938 | |
| 939 | /* stats.tx_window_errors*/ |
| 940 | new_stats->tx_window_errors = |
| 941 | amd8111e_read_mib(mmio, xmt_late_collision); |
| 942 | |
| 943 | /* Reset the mibs for collecting new statistics */ |
| 944 | /* writew(MIB_CLEAR, mmio + MIB_ADDR);*/ |
| 945 | |
| 946 | spin_unlock_irqrestore (&lp->lock, flags); |
| 947 | |
| 948 | return new_stats; |
| 949 | } |
| 950 | |
| 951 | /* This function recalculate the interrupt coalescing mode on every interrupt |
| 952 | * according to the datarate and the packet rate. |
| 953 | */ |
| 954 | static int amd8111e_calc_coalesce(struct net_device *dev) |
| 955 | { |
| 956 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 957 | struct amd8111e_coalesce_conf *coal_conf = &lp->coal_conf; |
| 958 | int tx_pkt_rate; |
| 959 | int rx_pkt_rate; |
| 960 | int tx_data_rate; |
| 961 | int rx_data_rate; |
| 962 | int rx_pkt_size; |
| 963 | int tx_pkt_size; |
| 964 | |
| 965 | tx_pkt_rate = coal_conf->tx_packets - coal_conf->tx_prev_packets; |
| 966 | coal_conf->tx_prev_packets = coal_conf->tx_packets; |
| 967 | |
| 968 | tx_data_rate = coal_conf->tx_bytes - coal_conf->tx_prev_bytes; |
| 969 | coal_conf->tx_prev_bytes = coal_conf->tx_bytes; |
| 970 | |
| 971 | rx_pkt_rate = coal_conf->rx_packets - coal_conf->rx_prev_packets; |
| 972 | coal_conf->rx_prev_packets = coal_conf->rx_packets; |
| 973 | |
| 974 | rx_data_rate = coal_conf->rx_bytes - coal_conf->rx_prev_bytes; |
| 975 | coal_conf->rx_prev_bytes = coal_conf->rx_bytes; |
| 976 | |
| 977 | if(rx_pkt_rate < 800){ |
| 978 | if(coal_conf->rx_coal_type != NO_COALESCE){ |
| 979 | |
| 980 | coal_conf->rx_timeout = 0x0; |
| 981 | coal_conf->rx_event_count = 0; |
| 982 | amd8111e_set_coalesce(dev,RX_INTR_COAL); |
| 983 | coal_conf->rx_coal_type = NO_COALESCE; |
| 984 | } |
| 985 | } |
| 986 | else{ |
| 987 | |
| 988 | rx_pkt_size = rx_data_rate/rx_pkt_rate; |
| 989 | if (rx_pkt_size < 128){ |
| 990 | if(coal_conf->rx_coal_type != NO_COALESCE){ |
| 991 | |
| 992 | coal_conf->rx_timeout = 0; |
| 993 | coal_conf->rx_event_count = 0; |
| 994 | amd8111e_set_coalesce(dev,RX_INTR_COAL); |
| 995 | coal_conf->rx_coal_type = NO_COALESCE; |
| 996 | } |
| 997 | |
| 998 | } |
| 999 | else if ( (rx_pkt_size >= 128) && (rx_pkt_size < 512) ){ |
| 1000 | |
| 1001 | if(coal_conf->rx_coal_type != LOW_COALESCE){ |
| 1002 | coal_conf->rx_timeout = 1; |
| 1003 | coal_conf->rx_event_count = 4; |
| 1004 | amd8111e_set_coalesce(dev,RX_INTR_COAL); |
| 1005 | coal_conf->rx_coal_type = LOW_COALESCE; |
| 1006 | } |
| 1007 | } |
| 1008 | else if ((rx_pkt_size >= 512) && (rx_pkt_size < 1024)){ |
| 1009 | |
| 1010 | if(coal_conf->rx_coal_type != MEDIUM_COALESCE){ |
| 1011 | coal_conf->rx_timeout = 1; |
| 1012 | coal_conf->rx_event_count = 4; |
| 1013 | amd8111e_set_coalesce(dev,RX_INTR_COAL); |
| 1014 | coal_conf->rx_coal_type = MEDIUM_COALESCE; |
| 1015 | } |
| 1016 | |
| 1017 | } |
| 1018 | else if(rx_pkt_size >= 1024){ |
| 1019 | if(coal_conf->rx_coal_type != HIGH_COALESCE){ |
| 1020 | coal_conf->rx_timeout = 2; |
| 1021 | coal_conf->rx_event_count = 3; |
| 1022 | amd8111e_set_coalesce(dev,RX_INTR_COAL); |
| 1023 | coal_conf->rx_coal_type = HIGH_COALESCE; |
| 1024 | } |
| 1025 | } |
| 1026 | } |
| 1027 | /* NOW FOR TX INTR COALESC */ |
| 1028 | if(tx_pkt_rate < 800){ |
| 1029 | if(coal_conf->tx_coal_type != NO_COALESCE){ |
| 1030 | |
| 1031 | coal_conf->tx_timeout = 0x0; |
| 1032 | coal_conf->tx_event_count = 0; |
| 1033 | amd8111e_set_coalesce(dev,TX_INTR_COAL); |
| 1034 | coal_conf->tx_coal_type = NO_COALESCE; |
| 1035 | } |
| 1036 | } |
| 1037 | else{ |
| 1038 | |
| 1039 | tx_pkt_size = tx_data_rate/tx_pkt_rate; |
| 1040 | if (tx_pkt_size < 128){ |
| 1041 | |
| 1042 | if(coal_conf->tx_coal_type != NO_COALESCE){ |
| 1043 | |
| 1044 | coal_conf->tx_timeout = 0; |
| 1045 | coal_conf->tx_event_count = 0; |
| 1046 | amd8111e_set_coalesce(dev,TX_INTR_COAL); |
| 1047 | coal_conf->tx_coal_type = NO_COALESCE; |
| 1048 | } |
| 1049 | |
| 1050 | } |
| 1051 | else if ( (tx_pkt_size >= 128) && (tx_pkt_size < 512) ){ |
| 1052 | |
| 1053 | if(coal_conf->tx_coal_type != LOW_COALESCE){ |
| 1054 | coal_conf->tx_timeout = 1; |
| 1055 | coal_conf->tx_event_count = 2; |
| 1056 | amd8111e_set_coalesce(dev,TX_INTR_COAL); |
| 1057 | coal_conf->tx_coal_type = LOW_COALESCE; |
| 1058 | |
| 1059 | } |
| 1060 | } |
| 1061 | else if ((tx_pkt_size >= 512) && (tx_pkt_size < 1024)){ |
| 1062 | |
| 1063 | if(coal_conf->tx_coal_type != MEDIUM_COALESCE){ |
| 1064 | coal_conf->tx_timeout = 2; |
| 1065 | coal_conf->tx_event_count = 5; |
| 1066 | amd8111e_set_coalesce(dev,TX_INTR_COAL); |
| 1067 | coal_conf->tx_coal_type = MEDIUM_COALESCE; |
| 1068 | } |
| 1069 | } else if (tx_pkt_size >= 1024) { |
| 1070 | if (coal_conf->tx_coal_type != HIGH_COALESCE) { |
| 1071 | coal_conf->tx_timeout = 4; |
| 1072 | coal_conf->tx_event_count = 8; |
| 1073 | amd8111e_set_coalesce(dev, TX_INTR_COAL); |
| 1074 | coal_conf->tx_coal_type = HIGH_COALESCE; |
| 1075 | } |
| 1076 | } |
| 1077 | } |
| 1078 | return 0; |
| 1079 | |
| 1080 | } |
| 1081 | |
| 1082 | /* This is device interrupt function. It handles transmit, |
| 1083 | * receive,link change and hardware timer interrupts. |
| 1084 | */ |
| 1085 | static irqreturn_t amd8111e_interrupt(int irq, void *dev_id) |
| 1086 | { |
| 1087 | |
| 1088 | struct net_device *dev = (struct net_device *)dev_id; |
| 1089 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1090 | void __iomem *mmio = lp->mmio; |
| 1091 | unsigned int intr0, intren0; |
| 1092 | unsigned int handled = 1; |
| 1093 | |
| 1094 | if(unlikely(dev == NULL)) |
| 1095 | return IRQ_NONE; |
| 1096 | |
| 1097 | spin_lock(&lp->lock); |
| 1098 | |
| 1099 | /* disabling interrupt */ |
| 1100 | writel(INTREN, mmio + CMD0); |
| 1101 | |
| 1102 | /* Read interrupt status */ |
| 1103 | intr0 = readl(mmio + INT0); |
| 1104 | intren0 = readl(mmio + INTEN0); |
| 1105 | |
| 1106 | /* Process all the INT event until INTR bit is clear. */ |
| 1107 | |
| 1108 | if (!(intr0 & INTR)){ |
| 1109 | handled = 0; |
| 1110 | goto err_no_interrupt; |
| 1111 | } |
| 1112 | |
| 1113 | /* Current driver processes 4 interrupts : RINT,TINT,LCINT,STINT */ |
| 1114 | writel(intr0, mmio + INT0); |
| 1115 | |
| 1116 | /* Check if Receive Interrupt has occurred. */ |
| 1117 | if (intr0 & RINT0) { |
| 1118 | if (napi_schedule_prep(&lp->napi)) { |
| 1119 | /* Disable receive interupts */ |
| 1120 | writel(RINTEN0, mmio + INTEN0); |
| 1121 | /* Schedule a polling routine */ |
| 1122 | __napi_schedule(&lp->napi); |
| 1123 | } else if (intren0 & RINTEN0) { |
| 1124 | netdev_dbg(dev, "************Driver bug! interrupt while in poll\n"); |
| 1125 | /* Fix by disable receive interrupts */ |
| 1126 | writel(RINTEN0, mmio + INTEN0); |
| 1127 | } |
| 1128 | } |
| 1129 | |
| 1130 | /* Check if Transmit Interrupt has occurred. */ |
| 1131 | if (intr0 & TINT0) |
| 1132 | amd8111e_tx(dev); |
| 1133 | |
| 1134 | /* Check if Link Change Interrupt has occurred. */ |
| 1135 | if (intr0 & LCINT) |
| 1136 | amd8111e_link_change(dev); |
| 1137 | |
| 1138 | /* Check if Hardware Timer Interrupt has occurred. */ |
| 1139 | if (intr0 & STINT) |
| 1140 | amd8111e_calc_coalesce(dev); |
| 1141 | |
| 1142 | err_no_interrupt: |
| 1143 | writel( VAL0 | INTREN,mmio + CMD0); |
| 1144 | |
| 1145 | spin_unlock(&lp->lock); |
| 1146 | |
| 1147 | return IRQ_RETVAL(handled); |
| 1148 | } |
| 1149 | |
| 1150 | #ifdef CONFIG_NET_POLL_CONTROLLER |
| 1151 | static void amd8111e_poll(struct net_device *dev) |
| 1152 | { |
| 1153 | unsigned long flags; |
| 1154 | local_irq_save(flags); |
| 1155 | amd8111e_interrupt(0, dev); |
| 1156 | local_irq_restore(flags); |
| 1157 | } |
| 1158 | #endif |
| 1159 | |
| 1160 | |
| 1161 | /* This function closes the network interface and updates |
| 1162 | * the statistics so that most recent statistics will be |
| 1163 | * available after the interface is down. |
| 1164 | */ |
| 1165 | static int amd8111e_close(struct net_device *dev) |
| 1166 | { |
| 1167 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1168 | netif_stop_queue(dev); |
| 1169 | |
| 1170 | napi_disable(&lp->napi); |
| 1171 | |
| 1172 | spin_lock_irq(&lp->lock); |
| 1173 | |
| 1174 | amd8111e_disable_interrupt(lp); |
| 1175 | amd8111e_stop_chip(lp); |
| 1176 | |
| 1177 | /* Free transmit and receive skbs */ |
| 1178 | amd8111e_free_skbs(lp->amd8111e_net_dev); |
| 1179 | |
| 1180 | netif_carrier_off(lp->amd8111e_net_dev); |
| 1181 | |
| 1182 | /* Delete ipg timer */ |
| 1183 | if(lp->options & OPTION_DYN_IPG_ENABLE) |
| 1184 | del_timer_sync(&lp->ipg_data.ipg_timer); |
| 1185 | |
| 1186 | spin_unlock_irq(&lp->lock); |
| 1187 | free_irq(dev->irq, dev); |
| 1188 | amd8111e_free_ring(lp); |
| 1189 | |
| 1190 | /* Update the statistics before closing */ |
| 1191 | amd8111e_get_stats(dev); |
| 1192 | lp->opened = 0; |
| 1193 | return 0; |
| 1194 | } |
| 1195 | |
| 1196 | /* This function opens new interface.It requests irq for the device, |
| 1197 | * initializes the device,buffers and descriptors, and starts the device. |
| 1198 | */ |
| 1199 | static int amd8111e_open(struct net_device *dev) |
| 1200 | { |
| 1201 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1202 | |
| 1203 | if(dev->irq ==0 || request_irq(dev->irq, amd8111e_interrupt, IRQF_SHARED, |
| 1204 | dev->name, dev)) |
| 1205 | return -EAGAIN; |
| 1206 | |
| 1207 | napi_enable(&lp->napi); |
| 1208 | |
| 1209 | spin_lock_irq(&lp->lock); |
| 1210 | |
| 1211 | amd8111e_init_hw_default(lp); |
| 1212 | |
| 1213 | if(amd8111e_restart(dev)){ |
| 1214 | spin_unlock_irq(&lp->lock); |
| 1215 | napi_disable(&lp->napi); |
| 1216 | if (dev->irq) |
| 1217 | free_irq(dev->irq, dev); |
| 1218 | return -ENOMEM; |
| 1219 | } |
| 1220 | /* Start ipg timer */ |
| 1221 | if(lp->options & OPTION_DYN_IPG_ENABLE){ |
| 1222 | add_timer(&lp->ipg_data.ipg_timer); |
| 1223 | netdev_info(dev, "Dynamic IPG Enabled\n"); |
| 1224 | } |
| 1225 | |
| 1226 | lp->opened = 1; |
| 1227 | |
| 1228 | spin_unlock_irq(&lp->lock); |
| 1229 | |
| 1230 | netif_start_queue(dev); |
| 1231 | |
| 1232 | return 0; |
| 1233 | } |
| 1234 | |
| 1235 | /* This function checks if there is any transmit descriptors |
| 1236 | * available to queue more packet. |
| 1237 | */ |
| 1238 | static int amd8111e_tx_queue_avail(struct amd8111e_priv *lp) |
| 1239 | { |
| 1240 | int tx_index = lp->tx_idx & TX_BUFF_MOD_MASK; |
| 1241 | if (lp->tx_skbuff[tx_index]) |
| 1242 | return -1; |
| 1243 | else |
| 1244 | return 0; |
| 1245 | |
| 1246 | } |
| 1247 | |
| 1248 | /* This function will queue the transmit packets to the |
| 1249 | * descriptors and will trigger the send operation. It also |
| 1250 | * initializes the transmit descriptors with buffer physical address, |
| 1251 | * byte count, ownership to hardware etc. |
| 1252 | */ |
| 1253 | static netdev_tx_t amd8111e_start_xmit(struct sk_buff *skb, |
| 1254 | struct net_device *dev) |
| 1255 | { |
| 1256 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1257 | int tx_index; |
| 1258 | unsigned long flags; |
| 1259 | |
| 1260 | spin_lock_irqsave(&lp->lock, flags); |
| 1261 | |
| 1262 | tx_index = lp->tx_idx & TX_RING_DR_MOD_MASK; |
| 1263 | |
| 1264 | lp->tx_ring[tx_index].buff_count = cpu_to_le16(skb->len); |
| 1265 | |
| 1266 | lp->tx_skbuff[tx_index] = skb; |
| 1267 | lp->tx_ring[tx_index].tx_flags = 0; |
| 1268 | |
| 1269 | #if AMD8111E_VLAN_TAG_USED |
| 1270 | if (skb_vlan_tag_present(skb)) { |
| 1271 | lp->tx_ring[tx_index].tag_ctrl_cmd |= |
| 1272 | cpu_to_le16(TCC_VLAN_INSERT); |
| 1273 | lp->tx_ring[tx_index].tag_ctrl_info = |
| 1274 | cpu_to_le16(skb_vlan_tag_get(skb)); |
| 1275 | |
| 1276 | } |
| 1277 | #endif |
| 1278 | lp->tx_dma_addr[tx_index] = |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1279 | dma_map_single(&lp->pci_dev->dev, skb->data, skb->len, |
| 1280 | DMA_TO_DEVICE); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1281 | lp->tx_ring[tx_index].buff_phy_addr = |
| 1282 | cpu_to_le32(lp->tx_dma_addr[tx_index]); |
| 1283 | |
| 1284 | /* Set FCS and LTINT bits */ |
| 1285 | wmb(); |
| 1286 | lp->tx_ring[tx_index].tx_flags |= |
| 1287 | cpu_to_le16(OWN_BIT | STP_BIT | ENP_BIT|ADD_FCS_BIT|LTINT_BIT); |
| 1288 | |
| 1289 | lp->tx_idx++; |
| 1290 | |
| 1291 | /* Trigger an immediate send poll. */ |
| 1292 | writel( VAL1 | TDMD0, lp->mmio + CMD0); |
| 1293 | writel( VAL2 | RDMD0,lp->mmio + CMD0); |
| 1294 | |
| 1295 | if(amd8111e_tx_queue_avail(lp) < 0){ |
| 1296 | netif_stop_queue(dev); |
| 1297 | } |
| 1298 | spin_unlock_irqrestore(&lp->lock, flags); |
| 1299 | return NETDEV_TX_OK; |
| 1300 | } |
| 1301 | /* This function returns all the memory mapped registers of the device. */ |
| 1302 | static void amd8111e_read_regs(struct amd8111e_priv *lp, u32 *buf) |
| 1303 | { |
| 1304 | void __iomem *mmio = lp->mmio; |
| 1305 | /* Read only necessary registers */ |
| 1306 | buf[0] = readl(mmio + XMT_RING_BASE_ADDR0); |
| 1307 | buf[1] = readl(mmio + XMT_RING_LEN0); |
| 1308 | buf[2] = readl(mmio + RCV_RING_BASE_ADDR0); |
| 1309 | buf[3] = readl(mmio + RCV_RING_LEN0); |
| 1310 | buf[4] = readl(mmio + CMD0); |
| 1311 | buf[5] = readl(mmio + CMD2); |
| 1312 | buf[6] = readl(mmio + CMD3); |
| 1313 | buf[7] = readl(mmio + CMD7); |
| 1314 | buf[8] = readl(mmio + INT0); |
| 1315 | buf[9] = readl(mmio + INTEN0); |
| 1316 | buf[10] = readl(mmio + LADRF); |
| 1317 | buf[11] = readl(mmio + LADRF+4); |
| 1318 | buf[12] = readl(mmio + STAT0); |
| 1319 | } |
| 1320 | |
| 1321 | |
| 1322 | /* This function sets promiscuos mode, all-multi mode or the multicast address |
| 1323 | * list to the device. |
| 1324 | */ |
| 1325 | static void amd8111e_set_multicast_list(struct net_device *dev) |
| 1326 | { |
| 1327 | struct netdev_hw_addr *ha; |
| 1328 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1329 | u32 mc_filter[2] ; |
| 1330 | int bit_num; |
| 1331 | |
| 1332 | if(dev->flags & IFF_PROMISC){ |
| 1333 | writel( VAL2 | PROM, lp->mmio + CMD2); |
| 1334 | return; |
| 1335 | } |
| 1336 | else |
| 1337 | writel( PROM, lp->mmio + CMD2); |
| 1338 | if (dev->flags & IFF_ALLMULTI || |
| 1339 | netdev_mc_count(dev) > MAX_FILTER_SIZE) { |
| 1340 | /* get all multicast packet */ |
| 1341 | mc_filter[1] = mc_filter[0] = 0xffffffff; |
| 1342 | lp->options |= OPTION_MULTICAST_ENABLE; |
| 1343 | amd8111e_writeq(*(u64 *)mc_filter, lp->mmio + LADRF); |
| 1344 | return; |
| 1345 | } |
| 1346 | if (netdev_mc_empty(dev)) { |
| 1347 | /* get only own packets */ |
| 1348 | mc_filter[1] = mc_filter[0] = 0; |
| 1349 | lp->options &= ~OPTION_MULTICAST_ENABLE; |
| 1350 | amd8111e_writeq(*(u64 *)mc_filter, lp->mmio + LADRF); |
| 1351 | /* disable promiscuous mode */ |
| 1352 | writel(PROM, lp->mmio + CMD2); |
| 1353 | return; |
| 1354 | } |
| 1355 | /* load all the multicast addresses in the logic filter */ |
| 1356 | lp->options |= OPTION_MULTICAST_ENABLE; |
| 1357 | mc_filter[1] = mc_filter[0] = 0; |
| 1358 | netdev_for_each_mc_addr(ha, dev) { |
| 1359 | bit_num = (ether_crc_le(ETH_ALEN, ha->addr) >> 26) & 0x3f; |
| 1360 | mc_filter[bit_num >> 5] |= 1 << (bit_num & 31); |
| 1361 | } |
| 1362 | amd8111e_writeq(*(u64 *)mc_filter, lp->mmio + LADRF); |
| 1363 | |
| 1364 | /* To eliminate PCI posting bug */ |
| 1365 | readl(lp->mmio + CMD2); |
| 1366 | |
| 1367 | } |
| 1368 | |
| 1369 | static void amd8111e_get_drvinfo(struct net_device *dev, |
| 1370 | struct ethtool_drvinfo *info) |
| 1371 | { |
| 1372 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1373 | struct pci_dev *pci_dev = lp->pci_dev; |
| 1374 | strlcpy(info->driver, MODULE_NAME, sizeof(info->driver)); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1375 | snprintf(info->fw_version, sizeof(info->fw_version), |
| 1376 | "%u", chip_version); |
| 1377 | strlcpy(info->bus_info, pci_name(pci_dev), sizeof(info->bus_info)); |
| 1378 | } |
| 1379 | |
| 1380 | static int amd8111e_get_regs_len(struct net_device *dev) |
| 1381 | { |
| 1382 | return AMD8111E_REG_DUMP_LEN; |
| 1383 | } |
| 1384 | |
| 1385 | static void amd8111e_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *buf) |
| 1386 | { |
| 1387 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1388 | regs->version = 0; |
| 1389 | amd8111e_read_regs(lp, buf); |
| 1390 | } |
| 1391 | |
| 1392 | static int amd8111e_get_link_ksettings(struct net_device *dev, |
| 1393 | struct ethtool_link_ksettings *cmd) |
| 1394 | { |
| 1395 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1396 | spin_lock_irq(&lp->lock); |
| 1397 | mii_ethtool_get_link_ksettings(&lp->mii_if, cmd); |
| 1398 | spin_unlock_irq(&lp->lock); |
| 1399 | return 0; |
| 1400 | } |
| 1401 | |
| 1402 | static int amd8111e_set_link_ksettings(struct net_device *dev, |
| 1403 | const struct ethtool_link_ksettings *cmd) |
| 1404 | { |
| 1405 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1406 | int res; |
| 1407 | spin_lock_irq(&lp->lock); |
| 1408 | res = mii_ethtool_set_link_ksettings(&lp->mii_if, cmd); |
| 1409 | spin_unlock_irq(&lp->lock); |
| 1410 | return res; |
| 1411 | } |
| 1412 | |
| 1413 | static int amd8111e_nway_reset(struct net_device *dev) |
| 1414 | { |
| 1415 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1416 | return mii_nway_restart(&lp->mii_if); |
| 1417 | } |
| 1418 | |
| 1419 | static u32 amd8111e_get_link(struct net_device *dev) |
| 1420 | { |
| 1421 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1422 | return mii_link_ok(&lp->mii_if); |
| 1423 | } |
| 1424 | |
| 1425 | static void amd8111e_get_wol(struct net_device *dev, struct ethtool_wolinfo *wol_info) |
| 1426 | { |
| 1427 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1428 | wol_info->supported = WAKE_MAGIC|WAKE_PHY; |
| 1429 | if (lp->options & OPTION_WOL_ENABLE) |
| 1430 | wol_info->wolopts = WAKE_MAGIC; |
| 1431 | } |
| 1432 | |
| 1433 | static int amd8111e_set_wol(struct net_device *dev, struct ethtool_wolinfo *wol_info) |
| 1434 | { |
| 1435 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1436 | if (wol_info->wolopts & ~(WAKE_MAGIC|WAKE_PHY)) |
| 1437 | return -EINVAL; |
| 1438 | spin_lock_irq(&lp->lock); |
| 1439 | if (wol_info->wolopts & WAKE_MAGIC) |
| 1440 | lp->options |= |
| 1441 | (OPTION_WOL_ENABLE | OPTION_WAKE_MAGIC_ENABLE); |
| 1442 | else if(wol_info->wolopts & WAKE_PHY) |
| 1443 | lp->options |= |
| 1444 | (OPTION_WOL_ENABLE | OPTION_WAKE_PHY_ENABLE); |
| 1445 | else |
| 1446 | lp->options &= ~OPTION_WOL_ENABLE; |
| 1447 | spin_unlock_irq(&lp->lock); |
| 1448 | return 0; |
| 1449 | } |
| 1450 | |
| 1451 | static const struct ethtool_ops ops = { |
| 1452 | .get_drvinfo = amd8111e_get_drvinfo, |
| 1453 | .get_regs_len = amd8111e_get_regs_len, |
| 1454 | .get_regs = amd8111e_get_regs, |
| 1455 | .nway_reset = amd8111e_nway_reset, |
| 1456 | .get_link = amd8111e_get_link, |
| 1457 | .get_wol = amd8111e_get_wol, |
| 1458 | .set_wol = amd8111e_set_wol, |
| 1459 | .get_link_ksettings = amd8111e_get_link_ksettings, |
| 1460 | .set_link_ksettings = amd8111e_set_link_ksettings, |
| 1461 | }; |
| 1462 | |
| 1463 | /* This function handles all the ethtool ioctls. It gives driver info, |
| 1464 | * gets/sets driver speed, gets memory mapped register values, forces |
| 1465 | * auto negotiation, sets/gets WOL options for ethtool application. |
| 1466 | */ |
| 1467 | static int amd8111e_ioctl(struct net_device *dev , struct ifreq *ifr, int cmd) |
| 1468 | { |
| 1469 | struct mii_ioctl_data *data = if_mii(ifr); |
| 1470 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1471 | int err; |
| 1472 | u32 mii_regval; |
| 1473 | |
| 1474 | switch(cmd) { |
| 1475 | case SIOCGMIIPHY: |
| 1476 | data->phy_id = lp->ext_phy_addr; |
| 1477 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1478 | fallthrough; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1479 | case SIOCGMIIREG: |
| 1480 | |
| 1481 | spin_lock_irq(&lp->lock); |
| 1482 | err = amd8111e_read_phy(lp, data->phy_id, |
| 1483 | data->reg_num & PHY_REG_ADDR_MASK, &mii_regval); |
| 1484 | spin_unlock_irq(&lp->lock); |
| 1485 | |
| 1486 | data->val_out = mii_regval; |
| 1487 | return err; |
| 1488 | |
| 1489 | case SIOCSMIIREG: |
| 1490 | |
| 1491 | spin_lock_irq(&lp->lock); |
| 1492 | err = amd8111e_write_phy(lp, data->phy_id, |
| 1493 | data->reg_num & PHY_REG_ADDR_MASK, data->val_in); |
| 1494 | spin_unlock_irq(&lp->lock); |
| 1495 | |
| 1496 | return err; |
| 1497 | |
| 1498 | default: |
| 1499 | /* do nothing */ |
| 1500 | break; |
| 1501 | } |
| 1502 | return -EOPNOTSUPP; |
| 1503 | } |
| 1504 | static int amd8111e_set_mac_address(struct net_device *dev, void *p) |
| 1505 | { |
| 1506 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1507 | int i; |
| 1508 | struct sockaddr *addr = p; |
| 1509 | |
| 1510 | memcpy(dev->dev_addr, addr->sa_data, dev->addr_len); |
| 1511 | spin_lock_irq(&lp->lock); |
| 1512 | /* Setting the MAC address to the device */ |
| 1513 | for (i = 0; i < ETH_ALEN; i++) |
| 1514 | writeb( dev->dev_addr[i], lp->mmio + PADR + i ); |
| 1515 | |
| 1516 | spin_unlock_irq(&lp->lock); |
| 1517 | |
| 1518 | return 0; |
| 1519 | } |
| 1520 | |
| 1521 | /* This function changes the mtu of the device. It restarts the device to |
| 1522 | * initialize the descriptor with new receive buffers. |
| 1523 | */ |
| 1524 | static int amd8111e_change_mtu(struct net_device *dev, int new_mtu) |
| 1525 | { |
| 1526 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1527 | int err; |
| 1528 | |
| 1529 | if (!netif_running(dev)) { |
| 1530 | /* new_mtu will be used |
| 1531 | * when device starts netxt time |
| 1532 | */ |
| 1533 | dev->mtu = new_mtu; |
| 1534 | return 0; |
| 1535 | } |
| 1536 | |
| 1537 | spin_lock_irq(&lp->lock); |
| 1538 | |
| 1539 | /* stop the chip */ |
| 1540 | writel(RUN, lp->mmio + CMD0); |
| 1541 | |
| 1542 | dev->mtu = new_mtu; |
| 1543 | |
| 1544 | err = amd8111e_restart(dev); |
| 1545 | spin_unlock_irq(&lp->lock); |
| 1546 | if(!err) |
| 1547 | netif_start_queue(dev); |
| 1548 | return err; |
| 1549 | } |
| 1550 | |
| 1551 | static int amd8111e_enable_magicpkt(struct amd8111e_priv *lp) |
| 1552 | { |
| 1553 | writel( VAL1|MPPLBA, lp->mmio + CMD3); |
| 1554 | writel( VAL0|MPEN_SW, lp->mmio + CMD7); |
| 1555 | |
| 1556 | /* To eliminate PCI posting bug */ |
| 1557 | readl(lp->mmio + CMD7); |
| 1558 | return 0; |
| 1559 | } |
| 1560 | |
| 1561 | static int amd8111e_enable_link_change(struct amd8111e_priv *lp) |
| 1562 | { |
| 1563 | |
| 1564 | /* Adapter is already stoped/suspended/interrupt-disabled */ |
| 1565 | writel(VAL0|LCMODE_SW,lp->mmio + CMD7); |
| 1566 | |
| 1567 | /* To eliminate PCI posting bug */ |
| 1568 | readl(lp->mmio + CMD7); |
| 1569 | return 0; |
| 1570 | } |
| 1571 | |
| 1572 | /* This function is called when a packet transmission fails to complete |
| 1573 | * within a reasonable period, on the assumption that an interrupt have |
| 1574 | * failed or the interface is locked up. This function will reinitialize |
| 1575 | * the hardware. |
| 1576 | */ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1577 | static void amd8111e_tx_timeout(struct net_device *dev, unsigned int txqueue) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1578 | { |
| 1579 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1580 | int err; |
| 1581 | |
| 1582 | netdev_err(dev, "transmit timed out, resetting\n"); |
| 1583 | |
| 1584 | spin_lock_irq(&lp->lock); |
| 1585 | err = amd8111e_restart(dev); |
| 1586 | spin_unlock_irq(&lp->lock); |
| 1587 | if(!err) |
| 1588 | netif_wake_queue(dev); |
| 1589 | } |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1590 | |
| 1591 | static int __maybe_unused amd8111e_suspend(struct device *dev_d) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1592 | { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1593 | struct net_device *dev = dev_get_drvdata(dev_d); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1594 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1595 | |
| 1596 | if (!netif_running(dev)) |
| 1597 | return 0; |
| 1598 | |
| 1599 | /* disable the interrupt */ |
| 1600 | spin_lock_irq(&lp->lock); |
| 1601 | amd8111e_disable_interrupt(lp); |
| 1602 | spin_unlock_irq(&lp->lock); |
| 1603 | |
| 1604 | netif_device_detach(dev); |
| 1605 | |
| 1606 | /* stop chip */ |
| 1607 | spin_lock_irq(&lp->lock); |
| 1608 | if(lp->options & OPTION_DYN_IPG_ENABLE) |
| 1609 | del_timer_sync(&lp->ipg_data.ipg_timer); |
| 1610 | amd8111e_stop_chip(lp); |
| 1611 | spin_unlock_irq(&lp->lock); |
| 1612 | |
| 1613 | if(lp->options & OPTION_WOL_ENABLE){ |
| 1614 | /* enable wol */ |
| 1615 | if(lp->options & OPTION_WAKE_MAGIC_ENABLE) |
| 1616 | amd8111e_enable_magicpkt(lp); |
| 1617 | if(lp->options & OPTION_WAKE_PHY_ENABLE) |
| 1618 | amd8111e_enable_link_change(lp); |
| 1619 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1620 | device_set_wakeup_enable(dev_d, 1); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1621 | |
| 1622 | } |
| 1623 | else{ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1624 | device_set_wakeup_enable(dev_d, 0); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1625 | } |
| 1626 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1627 | return 0; |
| 1628 | } |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1629 | |
| 1630 | static int __maybe_unused amd8111e_resume(struct device *dev_d) |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1631 | { |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1632 | struct net_device *dev = dev_get_drvdata(dev_d); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1633 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1634 | |
| 1635 | if (!netif_running(dev)) |
| 1636 | return 0; |
| 1637 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1638 | netif_device_attach(dev); |
| 1639 | |
| 1640 | spin_lock_irq(&lp->lock); |
| 1641 | amd8111e_restart(dev); |
| 1642 | /* Restart ipg timer */ |
| 1643 | if(lp->options & OPTION_DYN_IPG_ENABLE) |
| 1644 | mod_timer(&lp->ipg_data.ipg_timer, |
| 1645 | jiffies + IPG_CONVERGE_JIFFIES); |
| 1646 | spin_unlock_irq(&lp->lock); |
| 1647 | |
| 1648 | return 0; |
| 1649 | } |
| 1650 | |
| 1651 | static void amd8111e_config_ipg(struct timer_list *t) |
| 1652 | { |
| 1653 | struct amd8111e_priv *lp = from_timer(lp, t, ipg_data.ipg_timer); |
| 1654 | struct ipg_info *ipg_data = &lp->ipg_data; |
| 1655 | void __iomem *mmio = lp->mmio; |
| 1656 | unsigned int prev_col_cnt = ipg_data->col_cnt; |
| 1657 | unsigned int total_col_cnt; |
| 1658 | unsigned int tmp_ipg; |
| 1659 | |
| 1660 | if(lp->link_config.duplex == DUPLEX_FULL){ |
| 1661 | ipg_data->ipg = DEFAULT_IPG; |
| 1662 | return; |
| 1663 | } |
| 1664 | |
| 1665 | if(ipg_data->ipg_state == SSTATE){ |
| 1666 | |
| 1667 | if(ipg_data->timer_tick == IPG_STABLE_TIME){ |
| 1668 | |
| 1669 | ipg_data->timer_tick = 0; |
| 1670 | ipg_data->ipg = MIN_IPG - IPG_STEP; |
| 1671 | ipg_data->current_ipg = MIN_IPG; |
| 1672 | ipg_data->diff_col_cnt = 0xFFFFFFFF; |
| 1673 | ipg_data->ipg_state = CSTATE; |
| 1674 | } |
| 1675 | else |
| 1676 | ipg_data->timer_tick++; |
| 1677 | } |
| 1678 | |
| 1679 | if(ipg_data->ipg_state == CSTATE){ |
| 1680 | |
| 1681 | /* Get the current collision count */ |
| 1682 | |
| 1683 | total_col_cnt = ipg_data->col_cnt = |
| 1684 | amd8111e_read_mib(mmio, xmt_collisions); |
| 1685 | |
| 1686 | if ((total_col_cnt - prev_col_cnt) < |
| 1687 | (ipg_data->diff_col_cnt)){ |
| 1688 | |
| 1689 | ipg_data->diff_col_cnt = |
| 1690 | total_col_cnt - prev_col_cnt ; |
| 1691 | |
| 1692 | ipg_data->ipg = ipg_data->current_ipg; |
| 1693 | } |
| 1694 | |
| 1695 | ipg_data->current_ipg += IPG_STEP; |
| 1696 | |
| 1697 | if (ipg_data->current_ipg <= MAX_IPG) |
| 1698 | tmp_ipg = ipg_data->current_ipg; |
| 1699 | else{ |
| 1700 | tmp_ipg = ipg_data->ipg; |
| 1701 | ipg_data->ipg_state = SSTATE; |
| 1702 | } |
| 1703 | writew((u32)tmp_ipg, mmio + IPG); |
| 1704 | writew((u32)(tmp_ipg - IFS1_DELTA), mmio + IFS1); |
| 1705 | } |
David Brazdil | 0f672f6 | 2019-12-10 10:32:29 +0000 | [diff] [blame] | 1706 | mod_timer(&lp->ipg_data.ipg_timer, jiffies + IPG_CONVERGE_JIFFIES); |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1707 | return; |
| 1708 | |
| 1709 | } |
| 1710 | |
| 1711 | static void amd8111e_probe_ext_phy(struct net_device *dev) |
| 1712 | { |
| 1713 | struct amd8111e_priv *lp = netdev_priv(dev); |
| 1714 | int i; |
| 1715 | |
| 1716 | for (i = 0x1e; i >= 0; i--) { |
| 1717 | u32 id1, id2; |
| 1718 | |
| 1719 | if (amd8111e_read_phy(lp, i, MII_PHYSID1, &id1)) |
| 1720 | continue; |
| 1721 | if (amd8111e_read_phy(lp, i, MII_PHYSID2, &id2)) |
| 1722 | continue; |
| 1723 | lp->ext_phy_id = (id1 << 16) | id2; |
| 1724 | lp->ext_phy_addr = i; |
| 1725 | return; |
| 1726 | } |
| 1727 | lp->ext_phy_id = 0; |
| 1728 | lp->ext_phy_addr = 1; |
| 1729 | } |
| 1730 | |
| 1731 | static const struct net_device_ops amd8111e_netdev_ops = { |
| 1732 | .ndo_open = amd8111e_open, |
| 1733 | .ndo_stop = amd8111e_close, |
| 1734 | .ndo_start_xmit = amd8111e_start_xmit, |
| 1735 | .ndo_tx_timeout = amd8111e_tx_timeout, |
| 1736 | .ndo_get_stats = amd8111e_get_stats, |
| 1737 | .ndo_set_rx_mode = amd8111e_set_multicast_list, |
| 1738 | .ndo_validate_addr = eth_validate_addr, |
| 1739 | .ndo_set_mac_address = amd8111e_set_mac_address, |
| 1740 | .ndo_do_ioctl = amd8111e_ioctl, |
| 1741 | .ndo_change_mtu = amd8111e_change_mtu, |
| 1742 | #ifdef CONFIG_NET_POLL_CONTROLLER |
| 1743 | .ndo_poll_controller = amd8111e_poll, |
| 1744 | #endif |
| 1745 | }; |
| 1746 | |
| 1747 | static int amd8111e_probe_one(struct pci_dev *pdev, |
| 1748 | const struct pci_device_id *ent) |
| 1749 | { |
| 1750 | int err, i; |
| 1751 | unsigned long reg_addr,reg_len; |
| 1752 | struct amd8111e_priv *lp; |
| 1753 | struct net_device *dev; |
| 1754 | |
| 1755 | err = pci_enable_device(pdev); |
| 1756 | if(err){ |
| 1757 | dev_err(&pdev->dev, "Cannot enable new PCI device\n"); |
| 1758 | return err; |
| 1759 | } |
| 1760 | |
| 1761 | if(!(pci_resource_flags(pdev, 0) & IORESOURCE_MEM)){ |
| 1762 | dev_err(&pdev->dev, "Cannot find PCI base address\n"); |
| 1763 | err = -ENODEV; |
| 1764 | goto err_disable_pdev; |
| 1765 | } |
| 1766 | |
| 1767 | err = pci_request_regions(pdev, MODULE_NAME); |
| 1768 | if(err){ |
| 1769 | dev_err(&pdev->dev, "Cannot obtain PCI resources\n"); |
| 1770 | goto err_disable_pdev; |
| 1771 | } |
| 1772 | |
| 1773 | pci_set_master(pdev); |
| 1774 | |
| 1775 | /* Find power-management capability. */ |
| 1776 | if (!pdev->pm_cap) { |
| 1777 | dev_err(&pdev->dev, "No Power Management capability\n"); |
| 1778 | err = -ENODEV; |
| 1779 | goto err_free_reg; |
| 1780 | } |
| 1781 | |
| 1782 | /* Initialize DMA */ |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1783 | if (dma_set_mask(&pdev->dev, DMA_BIT_MASK(32)) < 0) { |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1784 | dev_err(&pdev->dev, "DMA not supported\n"); |
| 1785 | err = -ENODEV; |
| 1786 | goto err_free_reg; |
| 1787 | } |
| 1788 | |
| 1789 | reg_addr = pci_resource_start(pdev, 0); |
| 1790 | reg_len = pci_resource_len(pdev, 0); |
| 1791 | |
| 1792 | dev = alloc_etherdev(sizeof(struct amd8111e_priv)); |
| 1793 | if (!dev) { |
| 1794 | err = -ENOMEM; |
| 1795 | goto err_free_reg; |
| 1796 | } |
| 1797 | |
| 1798 | SET_NETDEV_DEV(dev, &pdev->dev); |
| 1799 | |
| 1800 | #if AMD8111E_VLAN_TAG_USED |
| 1801 | dev->features |= NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX ; |
| 1802 | #endif |
| 1803 | |
| 1804 | lp = netdev_priv(dev); |
| 1805 | lp->pci_dev = pdev; |
| 1806 | lp->amd8111e_net_dev = dev; |
| 1807 | lp->pm_cap = pdev->pm_cap; |
| 1808 | |
| 1809 | spin_lock_init(&lp->lock); |
| 1810 | |
| 1811 | lp->mmio = devm_ioremap(&pdev->dev, reg_addr, reg_len); |
| 1812 | if (!lp->mmio) { |
| 1813 | dev_err(&pdev->dev, "Cannot map device registers\n"); |
| 1814 | err = -ENOMEM; |
| 1815 | goto err_free_dev; |
| 1816 | } |
| 1817 | |
| 1818 | /* Initializing MAC address */ |
| 1819 | for (i = 0; i < ETH_ALEN; i++) |
| 1820 | dev->dev_addr[i] = readb(lp->mmio + PADR + i); |
| 1821 | |
| 1822 | /* Setting user defined parametrs */ |
| 1823 | lp->ext_phy_option = speed_duplex[card_idx]; |
| 1824 | if(coalesce[card_idx]) |
| 1825 | lp->options |= OPTION_INTR_COAL_ENABLE; |
| 1826 | if(dynamic_ipg[card_idx++]) |
| 1827 | lp->options |= OPTION_DYN_IPG_ENABLE; |
| 1828 | |
| 1829 | |
| 1830 | /* Initialize driver entry points */ |
| 1831 | dev->netdev_ops = &amd8111e_netdev_ops; |
| 1832 | dev->ethtool_ops = &ops; |
| 1833 | dev->irq =pdev->irq; |
| 1834 | dev->watchdog_timeo = AMD8111E_TX_TIMEOUT; |
| 1835 | dev->min_mtu = AMD8111E_MIN_MTU; |
| 1836 | dev->max_mtu = AMD8111E_MAX_MTU; |
| 1837 | netif_napi_add(dev, &lp->napi, amd8111e_rx_poll, 32); |
| 1838 | |
| 1839 | #if AMD8111E_VLAN_TAG_USED |
| 1840 | dev->features |= NETIF_F_HW_VLAN_CTAG_TX | NETIF_F_HW_VLAN_CTAG_RX; |
| 1841 | #endif |
| 1842 | /* Probe the external PHY */ |
| 1843 | amd8111e_probe_ext_phy(dev); |
| 1844 | |
| 1845 | /* setting mii default values */ |
| 1846 | lp->mii_if.dev = dev; |
| 1847 | lp->mii_if.mdio_read = amd8111e_mdio_read; |
| 1848 | lp->mii_if.mdio_write = amd8111e_mdio_write; |
| 1849 | lp->mii_if.phy_id = lp->ext_phy_addr; |
| 1850 | |
| 1851 | /* Set receive buffer length and set jumbo option*/ |
| 1852 | amd8111e_set_rx_buff_len(dev); |
| 1853 | |
| 1854 | |
| 1855 | err = register_netdev(dev); |
| 1856 | if (err) { |
| 1857 | dev_err(&pdev->dev, "Cannot register net device\n"); |
| 1858 | goto err_free_dev; |
| 1859 | } |
| 1860 | |
| 1861 | pci_set_drvdata(pdev, dev); |
| 1862 | |
| 1863 | /* Initialize software ipg timer */ |
| 1864 | if(lp->options & OPTION_DYN_IPG_ENABLE){ |
| 1865 | timer_setup(&lp->ipg_data.ipg_timer, amd8111e_config_ipg, 0); |
| 1866 | lp->ipg_data.ipg_timer.expires = jiffies + |
| 1867 | IPG_CONVERGE_JIFFIES; |
| 1868 | lp->ipg_data.ipg = DEFAULT_IPG; |
| 1869 | lp->ipg_data.ipg_state = CSTATE; |
| 1870 | } |
| 1871 | |
| 1872 | /* display driver and device information */ |
| 1873 | chip_version = (readl(lp->mmio + CHIPID) & 0xf0000000)>>28; |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1874 | dev_info(&pdev->dev, "[ Rev %x ] PCI 10/100BaseT Ethernet %pM\n", |
| 1875 | chip_version, dev->dev_addr); |
| 1876 | if (lp->ext_phy_id) |
| 1877 | dev_info(&pdev->dev, "Found MII PHY ID 0x%08x at address 0x%02x\n", |
| 1878 | lp->ext_phy_id, lp->ext_phy_addr); |
| 1879 | else |
| 1880 | dev_info(&pdev->dev, "Couldn't detect MII PHY, assuming address 0x01\n"); |
| 1881 | |
| 1882 | return 0; |
| 1883 | |
| 1884 | err_free_dev: |
| 1885 | free_netdev(dev); |
| 1886 | |
| 1887 | err_free_reg: |
| 1888 | pci_release_regions(pdev); |
| 1889 | |
| 1890 | err_disable_pdev: |
| 1891 | pci_disable_device(pdev); |
| 1892 | return err; |
| 1893 | |
| 1894 | } |
| 1895 | |
| 1896 | static void amd8111e_remove_one(struct pci_dev *pdev) |
| 1897 | { |
| 1898 | struct net_device *dev = pci_get_drvdata(pdev); |
| 1899 | |
| 1900 | if (dev) { |
| 1901 | unregister_netdev(dev); |
| 1902 | free_netdev(dev); |
| 1903 | pci_release_regions(pdev); |
| 1904 | pci_disable_device(pdev); |
| 1905 | } |
| 1906 | } |
| 1907 | |
| 1908 | static const struct pci_device_id amd8111e_pci_tbl[] = { |
| 1909 | { |
| 1910 | .vendor = PCI_VENDOR_ID_AMD, |
| 1911 | .device = PCI_DEVICE_ID_AMD8111E_7462, |
| 1912 | }, |
| 1913 | { |
| 1914 | .vendor = 0, |
| 1915 | } |
| 1916 | }; |
| 1917 | MODULE_DEVICE_TABLE(pci, amd8111e_pci_tbl); |
| 1918 | |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1919 | static SIMPLE_DEV_PM_OPS(amd8111e_pm_ops, amd8111e_suspend, amd8111e_resume); |
| 1920 | |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1921 | static struct pci_driver amd8111e_driver = { |
| 1922 | .name = MODULE_NAME, |
| 1923 | .id_table = amd8111e_pci_tbl, |
| 1924 | .probe = amd8111e_probe_one, |
| 1925 | .remove = amd8111e_remove_one, |
Olivier Deprez | 157378f | 2022-04-04 15:47:50 +0200 | [diff] [blame^] | 1926 | .driver.pm = &amd8111e_pm_ops |
Andrew Scull | b4b6d4a | 2019-01-02 15:54:55 +0000 | [diff] [blame] | 1927 | }; |
| 1928 | |
| 1929 | module_pci_driver(amd8111e_driver); |