1 | // SPDX-License-Identifier: GPL-2.0-or-later |
2 | /* |
3 | * Copyright(c) 2008 - 2009 Atheros Corporation. All rights reserved. |
4 | * |
5 | * Derived from Intel e1000 driver |
6 | * Copyright(c) 1999 - 2005 Intel Corporation. All rights reserved. |
7 | */ |
8 | |
9 | #include "atl1c.h" |
10 | |
11 | char atl1c_driver_name[] = "atl1c" ; |
12 | |
13 | /* |
14 | * atl1c_pci_tbl - PCI Device ID Table |
15 | * |
16 | * Wildcard entries (PCI_ANY_ID) should come last |
17 | * Last entry must be all 0s |
18 | * |
19 | * { Vendor ID, Device ID, SubVendor ID, SubDevice ID, |
20 | * Class, Class Mask, private data (not used) } |
21 | */ |
22 | static const struct pci_device_id atl1c_pci_tbl[] = { |
23 | {PCI_DEVICE(PCI_VENDOR_ID_ATTANSIC, PCI_DEVICE_ID_ATTANSIC_L1C)}, |
24 | {PCI_DEVICE(PCI_VENDOR_ID_ATTANSIC, PCI_DEVICE_ID_ATTANSIC_L2C)}, |
25 | {PCI_DEVICE(PCI_VENDOR_ID_ATTANSIC, PCI_DEVICE_ID_ATHEROS_L2C_B)}, |
26 | {PCI_DEVICE(PCI_VENDOR_ID_ATTANSIC, PCI_DEVICE_ID_ATHEROS_L2C_B2)}, |
27 | {PCI_DEVICE(PCI_VENDOR_ID_ATTANSIC, PCI_DEVICE_ID_ATHEROS_L1D)}, |
28 | {PCI_DEVICE(PCI_VENDOR_ID_ATTANSIC, PCI_DEVICE_ID_ATHEROS_L1D_2_0)}, |
29 | /* required last entry */ |
30 | { 0 } |
31 | }; |
32 | MODULE_DEVICE_TABLE(pci, atl1c_pci_tbl); |
33 | |
34 | MODULE_AUTHOR("Jie Yang" ); |
35 | MODULE_AUTHOR("Qualcomm Atheros Inc." ); |
36 | MODULE_DESCRIPTION("Qualcomm Atheros 100/1000M Ethernet Network Driver" ); |
37 | MODULE_LICENSE("GPL" ); |
38 | |
39 | struct atl1c_qregs { |
40 | u16 tpd_addr_lo; |
41 | u16 tpd_prod; |
42 | u16 tpd_cons; |
43 | u16 rfd_addr_lo; |
44 | u16 rrd_addr_lo; |
45 | u16 rfd_prod; |
46 | u32 tx_isr; |
47 | u32 rx_isr; |
48 | }; |
49 | |
50 | static struct atl1c_qregs atl1c_qregs[AT_MAX_TRANSMIT_QUEUE] = { |
51 | { |
52 | REG_TPD_PRI0_ADDR_LO, REG_TPD_PRI0_PIDX, REG_TPD_PRI0_CIDX, |
53 | REG_RFD0_HEAD_ADDR_LO, REG_RRD0_HEAD_ADDR_LO, |
54 | REG_MB_RFD0_PROD_IDX, ISR_TX_PKT_0, ISR_RX_PKT_0 |
55 | }, |
56 | { |
57 | REG_TPD_PRI1_ADDR_LO, REG_TPD_PRI1_PIDX, REG_TPD_PRI1_CIDX, |
58 | REG_RFD1_HEAD_ADDR_LO, REG_RRD1_HEAD_ADDR_LO, |
59 | REG_MB_RFD1_PROD_IDX, ISR_TX_PKT_1, ISR_RX_PKT_1 |
60 | }, |
61 | { |
62 | REG_TPD_PRI2_ADDR_LO, REG_TPD_PRI2_PIDX, REG_TPD_PRI2_CIDX, |
63 | REG_RFD2_HEAD_ADDR_LO, REG_RRD2_HEAD_ADDR_LO, |
64 | REG_MB_RFD2_PROD_IDX, ISR_TX_PKT_2, ISR_RX_PKT_2 |
65 | }, |
66 | { |
67 | REG_TPD_PRI3_ADDR_LO, REG_TPD_PRI3_PIDX, REG_TPD_PRI3_CIDX, |
68 | REG_RFD3_HEAD_ADDR_LO, REG_RRD3_HEAD_ADDR_LO, |
69 | REG_MB_RFD3_PROD_IDX, ISR_TX_PKT_3, ISR_RX_PKT_3 |
70 | }, |
71 | }; |
72 | |
73 | static int atl1c_stop_mac(struct atl1c_hw *hw); |
74 | static void atl1c_disable_l0s_l1(struct atl1c_hw *hw); |
75 | static void atl1c_set_aspm(struct atl1c_hw *hw, u16 link_speed); |
76 | static void atl1c_start_mac(struct atl1c_adapter *adapter); |
77 | static int atl1c_up(struct atl1c_adapter *adapter); |
78 | static void atl1c_down(struct atl1c_adapter *adapter); |
79 | static int atl1c_reset_mac(struct atl1c_hw *hw); |
80 | static void atl1c_reset_dma_ring(struct atl1c_adapter *adapter); |
81 | static int atl1c_configure(struct atl1c_adapter *adapter); |
82 | static int atl1c_alloc_rx_buffer(struct atl1c_adapter *adapter, u32 queue, |
83 | bool napi_mode); |
84 | |
85 | |
86 | static const u32 atl1c_default_msg = NETIF_MSG_DRV | NETIF_MSG_PROBE | |
87 | NETIF_MSG_LINK | NETIF_MSG_TIMER | NETIF_MSG_IFDOWN | NETIF_MSG_IFUP; |
88 | static void atl1c_pcie_patch(struct atl1c_hw *hw) |
89 | { |
90 | u32 mst_data, data; |
91 | |
92 | /* pclk sel could switch to 25M */ |
93 | AT_READ_REG(hw, REG_MASTER_CTRL, &mst_data); |
94 | mst_data &= ~MASTER_CTRL_CLK_SEL_DIS; |
95 | AT_WRITE_REG(hw, REG_MASTER_CTRL, mst_data); |
96 | |
97 | /* WoL/PCIE related settings */ |
98 | if (hw->nic_type == athr_l1c || hw->nic_type == athr_l2c) { |
99 | AT_READ_REG(hw, REG_PCIE_PHYMISC, &data); |
100 | data |= PCIE_PHYMISC_FORCE_RCV_DET; |
101 | AT_WRITE_REG(hw, REG_PCIE_PHYMISC, data); |
102 | } else { /* new dev set bit5 of MASTER */ |
103 | if (!(mst_data & MASTER_CTRL_WAKEN_25M)) |
104 | AT_WRITE_REG(hw, REG_MASTER_CTRL, |
105 | mst_data | MASTER_CTRL_WAKEN_25M); |
106 | } |
107 | /* aspm/PCIE setting only for l2cb 1.0 */ |
108 | if (hw->nic_type == athr_l2c_b && hw->revision_id == L2CB_V10) { |
109 | AT_READ_REG(hw, REG_PCIE_PHYMISC2, &data); |
110 | data = FIELD_SETX(data, PCIE_PHYMISC2_CDR_BW, |
111 | L2CB1_PCIE_PHYMISC2_CDR_BW); |
112 | data = FIELD_SETX(data, PCIE_PHYMISC2_L0S_TH, |
113 | L2CB1_PCIE_PHYMISC2_L0S_TH); |
114 | AT_WRITE_REG(hw, REG_PCIE_PHYMISC2, data); |
115 | /* extend L1 sync timer */ |
116 | AT_READ_REG(hw, REG_LINK_CTRL, &data); |
117 | data |= LINK_CTRL_EXT_SYNC; |
118 | AT_WRITE_REG(hw, REG_LINK_CTRL, data); |
119 | } |
120 | /* l2cb 1.x & l1d 1.x */ |
121 | if (hw->nic_type == athr_l2c_b || hw->nic_type == athr_l1d) { |
122 | AT_READ_REG(hw, REG_PM_CTRL, &data); |
123 | data |= PM_CTRL_L0S_BUFSRX_EN; |
124 | AT_WRITE_REG(hw, REG_PM_CTRL, data); |
125 | /* clear vendor msg */ |
126 | AT_READ_REG(hw, REG_DMA_DBG, &data); |
127 | AT_WRITE_REG(hw, REG_DMA_DBG, data & ~DMA_DBG_VENDOR_MSG); |
128 | } |
129 | } |
130 | |
131 | /* FIXME: no need any more ? */ |
132 | /* |
133 | * atl1c_init_pcie - init PCIE module |
134 | */ |
135 | static void atl1c_reset_pcie(struct atl1c_hw *hw, u32 flag) |
136 | { |
137 | u32 data; |
138 | u32 pci_cmd; |
139 | struct pci_dev *pdev = hw->adapter->pdev; |
140 | int pos; |
141 | |
142 | AT_READ_REG(hw, PCI_COMMAND, &pci_cmd); |
143 | pci_cmd &= ~PCI_COMMAND_INTX_DISABLE; |
144 | pci_cmd |= (PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER | |
145 | PCI_COMMAND_IO); |
146 | AT_WRITE_REG(hw, PCI_COMMAND, pci_cmd); |
147 | |
148 | /* |
149 | * Clear any PowerSaveing Settings |
150 | */ |
151 | pci_enable_wake(dev: pdev, PCI_D3hot, enable: 0); |
152 | pci_enable_wake(dev: pdev, PCI_D3cold, enable: 0); |
153 | /* wol sts read-clear */ |
154 | AT_READ_REG(hw, REG_WOL_CTRL, &data); |
155 | AT_WRITE_REG(hw, REG_WOL_CTRL, 0); |
156 | |
157 | /* |
158 | * Mask some pcie error bits |
159 | */ |
160 | pos = pci_find_ext_capability(dev: pdev, PCI_EXT_CAP_ID_ERR); |
161 | if (pos) { |
162 | pci_read_config_dword(dev: pdev, where: pos + PCI_ERR_UNCOR_SEVER, val: &data); |
163 | data &= ~(PCI_ERR_UNC_DLP | PCI_ERR_UNC_FCP); |
164 | pci_write_config_dword(dev: pdev, where: pos + PCI_ERR_UNCOR_SEVER, val: data); |
165 | } |
166 | /* clear error status */ |
167 | pcie_capability_write_word(dev: pdev, PCI_EXP_DEVSTA, |
168 | PCI_EXP_DEVSTA_NFED | |
169 | PCI_EXP_DEVSTA_FED | |
170 | PCI_EXP_DEVSTA_CED | |
171 | PCI_EXP_DEVSTA_URD); |
172 | |
173 | AT_READ_REG(hw, REG_LTSSM_ID_CTRL, &data); |
174 | data &= ~LTSSM_ID_EN_WRO; |
175 | AT_WRITE_REG(hw, REG_LTSSM_ID_CTRL, data); |
176 | |
177 | atl1c_pcie_patch(hw); |
178 | if (flag & ATL1C_PCIE_L0S_L1_DISABLE) |
179 | atl1c_disable_l0s_l1(hw); |
180 | |
181 | msleep(msecs: 5); |
182 | } |
183 | |
184 | /** |
185 | * atl1c_irq_enable - Enable default interrupt generation settings |
186 | * @adapter: board private structure |
187 | */ |
188 | static inline void atl1c_irq_enable(struct atl1c_adapter *adapter) |
189 | { |
190 | if (likely(atomic_dec_and_test(&adapter->irq_sem))) { |
191 | AT_WRITE_REG(&adapter->hw, REG_ISR, 0x7FFFFFFF); |
192 | AT_WRITE_REG(&adapter->hw, REG_IMR, adapter->hw.intr_mask); |
193 | AT_WRITE_FLUSH(&adapter->hw); |
194 | } |
195 | } |
196 | |
197 | /** |
198 | * atl1c_irq_disable - Mask off interrupt generation on the NIC |
199 | * @adapter: board private structure |
200 | */ |
201 | static inline void atl1c_irq_disable(struct atl1c_adapter *adapter) |
202 | { |
203 | atomic_inc(v: &adapter->irq_sem); |
204 | AT_WRITE_REG(&adapter->hw, REG_IMR, 0); |
205 | AT_WRITE_REG(&adapter->hw, REG_ISR, ISR_DIS_INT); |
206 | AT_WRITE_FLUSH(&adapter->hw); |
207 | synchronize_irq(irq: adapter->pdev->irq); |
208 | } |
209 | |
210 | /* |
211 | * atl1c_wait_until_idle - wait up to AT_HW_MAX_IDLE_DELAY reads |
212 | * of the idle status register until the device is actually idle |
213 | */ |
214 | static u32 atl1c_wait_until_idle(struct atl1c_hw *hw, u32 modu_ctrl) |
215 | { |
216 | int timeout; |
217 | u32 data; |
218 | |
219 | for (timeout = 0; timeout < AT_HW_MAX_IDLE_DELAY; timeout++) { |
220 | AT_READ_REG(hw, REG_IDLE_STATUS, &data); |
221 | if ((data & modu_ctrl) == 0) |
222 | return 0; |
223 | msleep(msecs: 1); |
224 | } |
225 | return data; |
226 | } |
227 | |
228 | /** |
229 | * atl1c_phy_config - Timer Call-back |
230 | * @t: timer list containing pointer to netdev cast into an unsigned long |
231 | */ |
232 | static void atl1c_phy_config(struct timer_list *t) |
233 | { |
234 | struct atl1c_adapter *adapter = from_timer(adapter, t, |
235 | phy_config_timer); |
236 | struct atl1c_hw *hw = &adapter->hw; |
237 | unsigned long flags; |
238 | |
239 | spin_lock_irqsave(&adapter->mdio_lock, flags); |
240 | atl1c_restart_autoneg(hw); |
241 | spin_unlock_irqrestore(lock: &adapter->mdio_lock, flags); |
242 | } |
243 | |
244 | void atl1c_reinit_locked(struct atl1c_adapter *adapter) |
245 | { |
246 | atl1c_down(adapter); |
247 | atl1c_up(adapter); |
248 | clear_bit(__AT_RESETTING, addr: &adapter->flags); |
249 | } |
250 | |
251 | static void atl1c_check_link_status(struct atl1c_adapter *adapter) |
252 | { |
253 | struct atl1c_hw *hw = &adapter->hw; |
254 | struct net_device *netdev = adapter->netdev; |
255 | struct pci_dev *pdev = adapter->pdev; |
256 | int err; |
257 | unsigned long flags; |
258 | u16 speed, duplex; |
259 | bool link; |
260 | |
261 | spin_lock_irqsave(&adapter->mdio_lock, flags); |
262 | link = atl1c_get_link_status(hw); |
263 | spin_unlock_irqrestore(lock: &adapter->mdio_lock, flags); |
264 | |
265 | if (!link) { |
266 | /* link down */ |
267 | netif_carrier_off(dev: netdev); |
268 | hw->hibernate = true; |
269 | if (atl1c_reset_mac(hw) != 0) |
270 | if (netif_msg_hw(adapter)) |
271 | dev_warn(&pdev->dev, "reset mac failed\n" ); |
272 | atl1c_set_aspm(hw, SPEED_0); |
273 | atl1c_post_phy_linkchg(hw, SPEED_0); |
274 | atl1c_reset_dma_ring(adapter); |
275 | atl1c_configure(adapter); |
276 | } else { |
277 | /* Link Up */ |
278 | hw->hibernate = false; |
279 | spin_lock_irqsave(&adapter->mdio_lock, flags); |
280 | err = atl1c_get_speed_and_duplex(hw, speed: &speed, duplex: &duplex); |
281 | spin_unlock_irqrestore(lock: &adapter->mdio_lock, flags); |
282 | if (unlikely(err)) |
283 | return; |
284 | /* link result is our setting */ |
285 | if (adapter->link_speed != speed || |
286 | adapter->link_duplex != duplex) { |
287 | adapter->link_speed = speed; |
288 | adapter->link_duplex = duplex; |
289 | atl1c_set_aspm(hw, link_speed: speed); |
290 | atl1c_post_phy_linkchg(hw, link_speed: speed); |
291 | atl1c_start_mac(adapter); |
292 | if (netif_msg_link(adapter)) |
293 | dev_info(&pdev->dev, |
294 | "%s: %s NIC Link is Up<%d Mbps %s>\n" , |
295 | atl1c_driver_name, netdev->name, |
296 | adapter->link_speed, |
297 | adapter->link_duplex == FULL_DUPLEX ? |
298 | "Full Duplex" : "Half Duplex" ); |
299 | } |
300 | if (!netif_carrier_ok(dev: netdev)) |
301 | netif_carrier_on(dev: netdev); |
302 | } |
303 | } |
304 | |
305 | static void atl1c_link_chg_event(struct atl1c_adapter *adapter) |
306 | { |
307 | struct net_device *netdev = adapter->netdev; |
308 | struct pci_dev *pdev = adapter->pdev; |
309 | bool link; |
310 | |
311 | spin_lock(lock: &adapter->mdio_lock); |
312 | link = atl1c_get_link_status(hw: &adapter->hw); |
313 | spin_unlock(lock: &adapter->mdio_lock); |
314 | /* notify upper layer link down ASAP */ |
315 | if (!link) { |
316 | if (netif_carrier_ok(dev: netdev)) { |
317 | /* old link state: Up */ |
318 | netif_carrier_off(dev: netdev); |
319 | if (netif_msg_link(adapter)) |
320 | dev_info(&pdev->dev, |
321 | "%s: %s NIC Link is Down\n" , |
322 | atl1c_driver_name, netdev->name); |
323 | adapter->link_speed = SPEED_0; |
324 | } |
325 | } |
326 | |
327 | set_bit(ATL1C_WORK_EVENT_LINK_CHANGE, addr: &adapter->work_event); |
328 | schedule_work(work: &adapter->common_task); |
329 | } |
330 | |
331 | static void atl1c_common_task(struct work_struct *work) |
332 | { |
333 | struct atl1c_adapter *adapter; |
334 | struct net_device *netdev; |
335 | |
336 | adapter = container_of(work, struct atl1c_adapter, common_task); |
337 | netdev = adapter->netdev; |
338 | |
339 | if (test_bit(__AT_DOWN, &adapter->flags)) |
340 | return; |
341 | |
342 | if (test_and_clear_bit(ATL1C_WORK_EVENT_RESET, addr: &adapter->work_event)) { |
343 | netif_device_detach(dev: netdev); |
344 | atl1c_down(adapter); |
345 | atl1c_up(adapter); |
346 | netif_device_attach(dev: netdev); |
347 | } |
348 | |
349 | if (test_and_clear_bit(ATL1C_WORK_EVENT_LINK_CHANGE, |
350 | addr: &adapter->work_event)) { |
351 | atl1c_irq_disable(adapter); |
352 | atl1c_check_link_status(adapter); |
353 | atl1c_irq_enable(adapter); |
354 | } |
355 | } |
356 | |
357 | |
358 | static void atl1c_del_timer(struct atl1c_adapter *adapter) |
359 | { |
360 | del_timer_sync(timer: &adapter->phy_config_timer); |
361 | } |
362 | |
363 | |
364 | /** |
365 | * atl1c_tx_timeout - Respond to a Tx Hang |
366 | * @netdev: network interface device structure |
367 | * @txqueue: index of hanging tx queue |
368 | */ |
369 | static void atl1c_tx_timeout(struct net_device *netdev, unsigned int txqueue) |
370 | { |
371 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
372 | |
373 | /* Do the reset outside of interrupt context */ |
374 | set_bit(ATL1C_WORK_EVENT_RESET, addr: &adapter->work_event); |
375 | schedule_work(work: &adapter->common_task); |
376 | } |
377 | |
378 | /** |
379 | * atl1c_set_multi - Multicast and Promiscuous mode set |
380 | * @netdev: network interface device structure |
381 | * |
382 | * The set_multi entry point is called whenever the multicast address |
383 | * list or the network interface flags are updated. This routine is |
384 | * responsible for configuring the hardware for proper multicast, |
385 | * promiscuous mode, and all-multi behavior. |
386 | */ |
387 | static void atl1c_set_multi(struct net_device *netdev) |
388 | { |
389 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
390 | struct atl1c_hw *hw = &adapter->hw; |
391 | struct netdev_hw_addr *ha; |
392 | u32 mac_ctrl_data; |
393 | u32 hash_value; |
394 | |
395 | /* Check for Promiscuous and All Multicast modes */ |
396 | AT_READ_REG(hw, REG_MAC_CTRL, &mac_ctrl_data); |
397 | |
398 | if (netdev->flags & IFF_PROMISC) { |
399 | mac_ctrl_data |= MAC_CTRL_PROMIS_EN; |
400 | } else if (netdev->flags & IFF_ALLMULTI) { |
401 | mac_ctrl_data |= MAC_CTRL_MC_ALL_EN; |
402 | mac_ctrl_data &= ~MAC_CTRL_PROMIS_EN; |
403 | } else { |
404 | mac_ctrl_data &= ~(MAC_CTRL_PROMIS_EN | MAC_CTRL_MC_ALL_EN); |
405 | } |
406 | |
407 | AT_WRITE_REG(hw, REG_MAC_CTRL, mac_ctrl_data); |
408 | |
409 | /* clear the old settings from the multicast hash table */ |
410 | AT_WRITE_REG(hw, REG_RX_HASH_TABLE, 0); |
411 | AT_WRITE_REG_ARRAY(hw, REG_RX_HASH_TABLE, 1, 0); |
412 | |
413 | /* comoute mc addresses' hash value ,and put it into hash table */ |
414 | netdev_for_each_mc_addr(ha, netdev) { |
415 | hash_value = atl1c_hash_mc_addr(hw, mc_addr: ha->addr); |
416 | atl1c_hash_set(hw, hash_value); |
417 | } |
418 | } |
419 | |
420 | static void __atl1c_vlan_mode(netdev_features_t features, u32 *mac_ctrl_data) |
421 | { |
422 | if (features & NETIF_F_HW_VLAN_CTAG_RX) { |
423 | /* enable VLAN tag insert/strip */ |
424 | *mac_ctrl_data |= MAC_CTRL_RMV_VLAN; |
425 | } else { |
426 | /* disable VLAN tag insert/strip */ |
427 | *mac_ctrl_data &= ~MAC_CTRL_RMV_VLAN; |
428 | } |
429 | } |
430 | |
431 | static void atl1c_vlan_mode(struct net_device *netdev, |
432 | netdev_features_t features) |
433 | { |
434 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
435 | struct pci_dev *pdev = adapter->pdev; |
436 | u32 mac_ctrl_data = 0; |
437 | |
438 | if (netif_msg_pktdata(adapter)) |
439 | dev_dbg(&pdev->dev, "atl1c_vlan_mode\n" ); |
440 | |
441 | atl1c_irq_disable(adapter); |
442 | AT_READ_REG(&adapter->hw, REG_MAC_CTRL, &mac_ctrl_data); |
443 | __atl1c_vlan_mode(features, mac_ctrl_data: &mac_ctrl_data); |
444 | AT_WRITE_REG(&adapter->hw, REG_MAC_CTRL, mac_ctrl_data); |
445 | atl1c_irq_enable(adapter); |
446 | } |
447 | |
448 | static void atl1c_restore_vlan(struct atl1c_adapter *adapter) |
449 | { |
450 | struct pci_dev *pdev = adapter->pdev; |
451 | |
452 | if (netif_msg_pktdata(adapter)) |
453 | dev_dbg(&pdev->dev, "atl1c_restore_vlan\n" ); |
454 | atl1c_vlan_mode(netdev: adapter->netdev, features: adapter->netdev->features); |
455 | } |
456 | |
457 | /** |
458 | * atl1c_set_mac_addr - Change the Ethernet Address of the NIC |
459 | * @netdev: network interface device structure |
460 | * @p: pointer to an address structure |
461 | * |
462 | * Returns 0 on success, negative on failure |
463 | */ |
464 | static int atl1c_set_mac_addr(struct net_device *netdev, void *p) |
465 | { |
466 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
467 | struct sockaddr *addr = p; |
468 | |
469 | if (!is_valid_ether_addr(addr: addr->sa_data)) |
470 | return -EADDRNOTAVAIL; |
471 | |
472 | if (netif_running(dev: netdev)) |
473 | return -EBUSY; |
474 | |
475 | eth_hw_addr_set(dev: netdev, addr: addr->sa_data); |
476 | memcpy(adapter->hw.mac_addr, addr->sa_data, netdev->addr_len); |
477 | |
478 | atl1c_hw_set_mac_addr(hw: &adapter->hw, mac_addr: adapter->hw.mac_addr); |
479 | |
480 | return 0; |
481 | } |
482 | |
483 | static void atl1c_set_rxbufsize(struct atl1c_adapter *adapter, |
484 | struct net_device *dev) |
485 | { |
486 | int mtu = dev->mtu; |
487 | |
488 | adapter->rx_buffer_len = mtu > AT_RX_BUF_SIZE ? |
489 | roundup(mtu + ETH_HLEN + ETH_FCS_LEN + VLAN_HLEN, 8) : AT_RX_BUF_SIZE; |
490 | } |
491 | |
492 | static netdev_features_t atl1c_fix_features(struct net_device *netdev, |
493 | netdev_features_t features) |
494 | { |
495 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
496 | struct atl1c_hw *hw = &adapter->hw; |
497 | |
498 | /* |
499 | * Since there is no support for separate rx/tx vlan accel |
500 | * enable/disable make sure tx flag is always in same state as rx. |
501 | */ |
502 | if (features & NETIF_F_HW_VLAN_CTAG_RX) |
503 | features |= NETIF_F_HW_VLAN_CTAG_TX; |
504 | else |
505 | features &= ~NETIF_F_HW_VLAN_CTAG_TX; |
506 | |
507 | if (hw->nic_type != athr_mt) { |
508 | if (netdev->mtu > MAX_TSO_FRAME_SIZE) |
509 | features &= ~(NETIF_F_TSO | NETIF_F_TSO6); |
510 | } |
511 | |
512 | return features; |
513 | } |
514 | |
515 | static int atl1c_set_features(struct net_device *netdev, |
516 | netdev_features_t features) |
517 | { |
518 | netdev_features_t changed = netdev->features ^ features; |
519 | |
520 | if (changed & NETIF_F_HW_VLAN_CTAG_RX) |
521 | atl1c_vlan_mode(netdev, features); |
522 | |
523 | return 0; |
524 | } |
525 | |
526 | static void atl1c_set_max_mtu(struct net_device *netdev) |
527 | { |
528 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
529 | struct atl1c_hw *hw = &adapter->hw; |
530 | |
531 | switch (hw->nic_type) { |
532 | /* These (GbE) devices support jumbo packets, max_mtu 6122 */ |
533 | case athr_l1c: |
534 | case athr_l1d: |
535 | case athr_l1d_2: |
536 | netdev->max_mtu = MAX_JUMBO_FRAME_SIZE - |
537 | (ETH_HLEN + ETH_FCS_LEN + VLAN_HLEN); |
538 | break; |
539 | case athr_mt: |
540 | netdev->max_mtu = 9500; |
541 | break; |
542 | /* The 10/100 devices don't support jumbo packets, max_mtu 1500 */ |
543 | default: |
544 | netdev->max_mtu = ETH_DATA_LEN; |
545 | break; |
546 | } |
547 | } |
548 | |
549 | /** |
550 | * atl1c_change_mtu - Change the Maximum Transfer Unit |
551 | * @netdev: network interface device structure |
552 | * @new_mtu: new value for maximum frame size |
553 | * |
554 | * Returns 0 on success, negative on failure |
555 | */ |
556 | static int atl1c_change_mtu(struct net_device *netdev, int new_mtu) |
557 | { |
558 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
559 | |
560 | /* set MTU */ |
561 | if (netif_running(dev: netdev)) { |
562 | while (test_and_set_bit(__AT_RESETTING, addr: &adapter->flags)) |
563 | msleep(msecs: 1); |
564 | netdev->mtu = new_mtu; |
565 | adapter->hw.max_frame_size = new_mtu; |
566 | atl1c_set_rxbufsize(adapter, dev: netdev); |
567 | atl1c_down(adapter); |
568 | netdev_update_features(dev: netdev); |
569 | atl1c_up(adapter); |
570 | clear_bit(__AT_RESETTING, addr: &adapter->flags); |
571 | } |
572 | return 0; |
573 | } |
574 | |
575 | /* |
576 | * caller should hold mdio_lock |
577 | */ |
578 | static int atl1c_mdio_read(struct net_device *netdev, int phy_id, int reg_num) |
579 | { |
580 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
581 | u16 result; |
582 | |
583 | atl1c_read_phy_reg(hw: &adapter->hw, reg_addr: reg_num, phy_data: &result); |
584 | return result; |
585 | } |
586 | |
587 | static void atl1c_mdio_write(struct net_device *netdev, int phy_id, |
588 | int reg_num, int val) |
589 | { |
590 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
591 | |
592 | atl1c_write_phy_reg(hw: &adapter->hw, reg_addr: reg_num, phy_data: val); |
593 | } |
594 | |
595 | static int atl1c_mii_ioctl(struct net_device *netdev, |
596 | struct ifreq *ifr, int cmd) |
597 | { |
598 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
599 | struct pci_dev *pdev = adapter->pdev; |
600 | struct mii_ioctl_data *data = if_mii(rq: ifr); |
601 | unsigned long flags; |
602 | int retval = 0; |
603 | |
604 | if (!netif_running(dev: netdev)) |
605 | return -EINVAL; |
606 | |
607 | spin_lock_irqsave(&adapter->mdio_lock, flags); |
608 | switch (cmd) { |
609 | case SIOCGMIIPHY: |
610 | data->phy_id = 0; |
611 | break; |
612 | |
613 | case SIOCGMIIREG: |
614 | if (atl1c_read_phy_reg(hw: &adapter->hw, reg_addr: data->reg_num & 0x1F, |
615 | phy_data: &data->val_out)) { |
616 | retval = -EIO; |
617 | goto out; |
618 | } |
619 | break; |
620 | |
621 | case SIOCSMIIREG: |
622 | if (data->reg_num & ~(0x1F)) { |
623 | retval = -EFAULT; |
624 | goto out; |
625 | } |
626 | |
627 | dev_dbg(&pdev->dev, "<atl1c_mii_ioctl> write %x %x" , |
628 | data->reg_num, data->val_in); |
629 | if (atl1c_write_phy_reg(hw: &adapter->hw, |
630 | reg_addr: data->reg_num, phy_data: data->val_in)) { |
631 | retval = -EIO; |
632 | goto out; |
633 | } |
634 | break; |
635 | |
636 | default: |
637 | retval = -EOPNOTSUPP; |
638 | break; |
639 | } |
640 | out: |
641 | spin_unlock_irqrestore(lock: &adapter->mdio_lock, flags); |
642 | return retval; |
643 | } |
644 | |
645 | static int atl1c_ioctl(struct net_device *netdev, struct ifreq *ifr, int cmd) |
646 | { |
647 | switch (cmd) { |
648 | case SIOCGMIIPHY: |
649 | case SIOCGMIIREG: |
650 | case SIOCSMIIREG: |
651 | return atl1c_mii_ioctl(netdev, ifr, cmd); |
652 | default: |
653 | return -EOPNOTSUPP; |
654 | } |
655 | } |
656 | |
657 | /** |
658 | * atl1c_alloc_queues - Allocate memory for all rings |
659 | * @adapter: board private structure to initialize |
660 | * |
661 | */ |
662 | static int atl1c_alloc_queues(struct atl1c_adapter *adapter) |
663 | { |
664 | return 0; |
665 | } |
666 | |
667 | static enum atl1c_nic_type atl1c_get_mac_type(struct pci_dev *pdev, |
668 | u8 __iomem *hw_addr) |
669 | { |
670 | switch (pdev->device) { |
671 | case PCI_DEVICE_ID_ATTANSIC_L2C: |
672 | return athr_l2c; |
673 | case PCI_DEVICE_ID_ATTANSIC_L1C: |
674 | return athr_l1c; |
675 | case PCI_DEVICE_ID_ATHEROS_L2C_B: |
676 | return athr_l2c_b; |
677 | case PCI_DEVICE_ID_ATHEROS_L2C_B2: |
678 | return athr_l2c_b2; |
679 | case PCI_DEVICE_ID_ATHEROS_L1D: |
680 | return athr_l1d; |
681 | case PCI_DEVICE_ID_ATHEROS_L1D_2_0: |
682 | if (readl(addr: hw_addr + REG_MT_MAGIC) == MT_MAGIC) |
683 | return athr_mt; |
684 | return athr_l1d_2; |
685 | default: |
686 | return athr_l1c; |
687 | } |
688 | } |
689 | |
690 | static int atl1c_setup_mac_funcs(struct atl1c_hw *hw) |
691 | { |
692 | u32 link_ctrl_data; |
693 | |
694 | AT_READ_REG(hw, REG_LINK_CTRL, &link_ctrl_data); |
695 | |
696 | hw->ctrl_flags = ATL1C_INTR_MODRT_ENABLE | |
697 | ATL1C_TXQ_MODE_ENHANCE; |
698 | hw->ctrl_flags |= ATL1C_ASPM_L0S_SUPPORT | |
699 | ATL1C_ASPM_L1_SUPPORT; |
700 | hw->ctrl_flags |= ATL1C_ASPM_CTRL_MON; |
701 | |
702 | if (hw->nic_type == athr_l1c || |
703 | hw->nic_type == athr_l1d || |
704 | hw->nic_type == athr_l1d_2) |
705 | hw->link_cap_flags |= ATL1C_LINK_CAP_1000M; |
706 | return 0; |
707 | } |
708 | |
709 | struct atl1c_platform_patch { |
710 | u16 pci_did; |
711 | u8 pci_revid; |
712 | u16 subsystem_vid; |
713 | u16 subsystem_did; |
714 | u32 patch_flag; |
715 | #define ATL1C_LINK_PATCH 0x1 |
716 | }; |
717 | static const struct atl1c_platform_patch plats[] = { |
718 | {0x2060, 0xC1, 0x1019, 0x8152, 0x1}, |
719 | {0x2060, 0xC1, 0x1019, 0x2060, 0x1}, |
720 | {0x2060, 0xC1, 0x1019, 0xE000, 0x1}, |
721 | {0x2062, 0xC0, 0x1019, 0x8152, 0x1}, |
722 | {0x2062, 0xC0, 0x1019, 0x2062, 0x1}, |
723 | {0x2062, 0xC0, 0x1458, 0xE000, 0x1}, |
724 | {0x2062, 0xC1, 0x1019, 0x8152, 0x1}, |
725 | {0x2062, 0xC1, 0x1019, 0x2062, 0x1}, |
726 | {0x2062, 0xC1, 0x1458, 0xE000, 0x1}, |
727 | {0x2062, 0xC1, 0x1565, 0x2802, 0x1}, |
728 | {0x2062, 0xC1, 0x1565, 0x2801, 0x1}, |
729 | {0x1073, 0xC0, 0x1019, 0x8151, 0x1}, |
730 | {0x1073, 0xC0, 0x1019, 0x1073, 0x1}, |
731 | {0x1073, 0xC0, 0x1458, 0xE000, 0x1}, |
732 | {0x1083, 0xC0, 0x1458, 0xE000, 0x1}, |
733 | {0x1083, 0xC0, 0x1019, 0x8151, 0x1}, |
734 | {0x1083, 0xC0, 0x1019, 0x1083, 0x1}, |
735 | {0x1083, 0xC0, 0x1462, 0x7680, 0x1}, |
736 | {0x1083, 0xC0, 0x1565, 0x2803, 0x1}, |
737 | {0}, |
738 | }; |
739 | |
740 | static void atl1c_patch_assign(struct atl1c_hw *hw) |
741 | { |
742 | struct pci_dev *pdev = hw->adapter->pdev; |
743 | u32 misc_ctrl; |
744 | int i = 0; |
745 | |
746 | hw->msi_lnkpatch = false; |
747 | |
748 | while (plats[i].pci_did != 0) { |
749 | if (plats[i].pci_did == hw->device_id && |
750 | plats[i].pci_revid == hw->revision_id && |
751 | plats[i].subsystem_vid == hw->subsystem_vendor_id && |
752 | plats[i].subsystem_did == hw->subsystem_id) { |
753 | if (plats[i].patch_flag & ATL1C_LINK_PATCH) |
754 | hw->msi_lnkpatch = true; |
755 | } |
756 | i++; |
757 | } |
758 | |
759 | if (hw->device_id == PCI_DEVICE_ID_ATHEROS_L2C_B2 && |
760 | hw->revision_id == L2CB_V21) { |
761 | /* config access mode */ |
762 | pci_write_config_dword(dev: pdev, REG_PCIE_IND_ACC_ADDR, |
763 | REG_PCIE_DEV_MISC_CTRL); |
764 | pci_read_config_dword(dev: pdev, REG_PCIE_IND_ACC_DATA, val: &misc_ctrl); |
765 | misc_ctrl &= ~0x100; |
766 | pci_write_config_dword(dev: pdev, REG_PCIE_IND_ACC_ADDR, |
767 | REG_PCIE_DEV_MISC_CTRL); |
768 | pci_write_config_dword(dev: pdev, REG_PCIE_IND_ACC_DATA, val: misc_ctrl); |
769 | } |
770 | } |
771 | /** |
772 | * atl1c_sw_init - Initialize general software structures (struct atl1c_adapter) |
773 | * @adapter: board private structure to initialize |
774 | * |
775 | * atl1c_sw_init initializes the Adapter private data structure. |
776 | * Fields are initialized based on PCI device information and |
777 | * OS network device settings (MTU size). |
778 | */ |
779 | static int atl1c_sw_init(struct atl1c_adapter *adapter) |
780 | { |
781 | struct atl1c_hw *hw = &adapter->hw; |
782 | struct pci_dev *pdev = adapter->pdev; |
783 | u32 revision; |
784 | int i; |
785 | |
786 | adapter->wol = 0; |
787 | device_set_wakeup_enable(dev: &pdev->dev, enable: false); |
788 | adapter->link_speed = SPEED_0; |
789 | adapter->link_duplex = FULL_DUPLEX; |
790 | adapter->tpd_ring[0].count = 1024; |
791 | adapter->rfd_ring[0].count = 512; |
792 | |
793 | hw->vendor_id = pdev->vendor; |
794 | hw->device_id = pdev->device; |
795 | hw->subsystem_vendor_id = pdev->subsystem_vendor; |
796 | hw->subsystem_id = pdev->subsystem_device; |
797 | pci_read_config_dword(dev: pdev, PCI_CLASS_REVISION, val: &revision); |
798 | hw->revision_id = revision & 0xFF; |
799 | /* before link up, we assume hibernate is true */ |
800 | hw->hibernate = true; |
801 | hw->media_type = MEDIA_TYPE_AUTO_SENSOR; |
802 | if (atl1c_setup_mac_funcs(hw) != 0) { |
803 | dev_err(&pdev->dev, "set mac function pointers failed\n" ); |
804 | return -1; |
805 | } |
806 | atl1c_patch_assign(hw); |
807 | |
808 | hw->intr_mask = IMR_NORMAL_MASK; |
809 | for (i = 0; i < adapter->tx_queue_count; ++i) |
810 | hw->intr_mask |= atl1c_qregs[i].tx_isr; |
811 | for (i = 0; i < adapter->rx_queue_count; ++i) |
812 | hw->intr_mask |= atl1c_qregs[i].rx_isr; |
813 | hw->phy_configured = false; |
814 | hw->preamble_len = 7; |
815 | hw->max_frame_size = adapter->netdev->mtu; |
816 | hw->autoneg_advertised = ADVERTISED_Autoneg; |
817 | hw->indirect_tab = 0xE4E4E4E4; |
818 | hw->base_cpu = 0; |
819 | |
820 | hw->ict = 50000; /* 100ms */ |
821 | hw->smb_timer = 200000; /* 400ms */ |
822 | hw->rx_imt = 200; |
823 | hw->tx_imt = 1000; |
824 | |
825 | hw->tpd_burst = 5; |
826 | hw->rfd_burst = 8; |
827 | hw->dma_order = atl1c_dma_ord_out; |
828 | hw->dmar_block = atl1c_dma_req_1024; |
829 | |
830 | if (atl1c_alloc_queues(adapter)) { |
831 | dev_err(&pdev->dev, "Unable to allocate memory for queues\n" ); |
832 | return -ENOMEM; |
833 | } |
834 | /* TODO */ |
835 | atl1c_set_rxbufsize(adapter, dev: adapter->netdev); |
836 | atomic_set(v: &adapter->irq_sem, i: 1); |
837 | spin_lock_init(&adapter->mdio_lock); |
838 | spin_lock_init(&adapter->hw.intr_mask_lock); |
839 | set_bit(__AT_DOWN, addr: &adapter->flags); |
840 | |
841 | return 0; |
842 | } |
843 | |
844 | static inline void atl1c_clean_buffer(struct pci_dev *pdev, |
845 | struct atl1c_buffer *buffer_info, |
846 | int budget) |
847 | { |
848 | u16 pci_driection; |
849 | if (buffer_info->flags & ATL1C_BUFFER_FREE) |
850 | return; |
851 | if (buffer_info->dma) { |
852 | if (buffer_info->flags & ATL1C_PCIMAP_FROMDEVICE) |
853 | pci_driection = DMA_FROM_DEVICE; |
854 | else |
855 | pci_driection = DMA_TO_DEVICE; |
856 | |
857 | if (buffer_info->flags & ATL1C_PCIMAP_SINGLE) |
858 | dma_unmap_single(&pdev->dev, buffer_info->dma, |
859 | buffer_info->length, pci_driection); |
860 | else if (buffer_info->flags & ATL1C_PCIMAP_PAGE) |
861 | dma_unmap_page(&pdev->dev, buffer_info->dma, |
862 | buffer_info->length, pci_driection); |
863 | } |
864 | if (buffer_info->skb) |
865 | napi_consume_skb(skb: buffer_info->skb, budget); |
866 | buffer_info->dma = 0; |
867 | buffer_info->skb = NULL; |
868 | ATL1C_SET_BUFFER_STATE(buffer_info, ATL1C_BUFFER_FREE); |
869 | } |
870 | /** |
871 | * atl1c_clean_tx_ring - Free Tx-skb |
872 | * @adapter: board private structure |
873 | * @queue: idx of transmit queue |
874 | */ |
875 | static void atl1c_clean_tx_ring(struct atl1c_adapter *adapter, |
876 | u32 queue) |
877 | { |
878 | struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue]; |
879 | struct atl1c_buffer *buffer_info; |
880 | struct pci_dev *pdev = adapter->pdev; |
881 | u16 index, ring_count; |
882 | |
883 | ring_count = tpd_ring->count; |
884 | for (index = 0; index < ring_count; index++) { |
885 | buffer_info = &tpd_ring->buffer_info[index]; |
886 | atl1c_clean_buffer(pdev, buffer_info, budget: 0); |
887 | } |
888 | |
889 | netdev_tx_reset_queue(q: netdev_get_tx_queue(dev: adapter->netdev, index: queue)); |
890 | |
891 | /* Zero out Tx-buffers */ |
892 | memset(tpd_ring->desc, 0, sizeof(struct atl1c_tpd_desc) * |
893 | ring_count); |
894 | atomic_set(v: &tpd_ring->next_to_clean, i: 0); |
895 | tpd_ring->next_to_use = 0; |
896 | } |
897 | |
898 | /** |
899 | * atl1c_clean_rx_ring - Free rx-reservation skbs |
900 | * @adapter: board private structure |
901 | * @queue: idx of transmit queue |
902 | */ |
903 | static void atl1c_clean_rx_ring(struct atl1c_adapter *adapter, u32 queue) |
904 | { |
905 | struct atl1c_rfd_ring *rfd_ring = &adapter->rfd_ring[queue]; |
906 | struct atl1c_rrd_ring *rrd_ring = &adapter->rrd_ring[queue]; |
907 | struct atl1c_buffer *buffer_info; |
908 | struct pci_dev *pdev = adapter->pdev; |
909 | int j; |
910 | |
911 | for (j = 0; j < rfd_ring->count; j++) { |
912 | buffer_info = &rfd_ring->buffer_info[j]; |
913 | atl1c_clean_buffer(pdev, buffer_info, budget: 0); |
914 | } |
915 | /* zero out the descriptor ring */ |
916 | memset(rfd_ring->desc, 0, rfd_ring->size); |
917 | rfd_ring->next_to_clean = 0; |
918 | rfd_ring->next_to_use = 0; |
919 | rrd_ring->next_to_use = 0; |
920 | rrd_ring->next_to_clean = 0; |
921 | } |
922 | |
923 | /* |
924 | * Read / Write Ptr Initialize: |
925 | */ |
926 | static void atl1c_init_ring_ptrs(struct atl1c_adapter *adapter) |
927 | { |
928 | struct atl1c_tpd_ring *tpd_ring = adapter->tpd_ring; |
929 | struct atl1c_rfd_ring *rfd_ring = adapter->rfd_ring; |
930 | struct atl1c_rrd_ring *rrd_ring = adapter->rrd_ring; |
931 | struct atl1c_buffer *buffer_info; |
932 | int i, j; |
933 | |
934 | for (i = 0; i < adapter->tx_queue_count; i++) { |
935 | tpd_ring[i].next_to_use = 0; |
936 | atomic_set(v: &tpd_ring[i].next_to_clean, i: 0); |
937 | buffer_info = tpd_ring[i].buffer_info; |
938 | for (j = 0; j < tpd_ring->count; j++) |
939 | ATL1C_SET_BUFFER_STATE(&buffer_info[i], |
940 | ATL1C_BUFFER_FREE); |
941 | } |
942 | for (i = 0; i < adapter->rx_queue_count; i++) { |
943 | rfd_ring[i].next_to_use = 0; |
944 | rfd_ring[i].next_to_clean = 0; |
945 | rrd_ring[i].next_to_use = 0; |
946 | rrd_ring[i].next_to_clean = 0; |
947 | for (j = 0; j < rfd_ring[i].count; j++) { |
948 | buffer_info = &rfd_ring[i].buffer_info[j]; |
949 | ATL1C_SET_BUFFER_STATE(buffer_info, ATL1C_BUFFER_FREE); |
950 | } |
951 | } |
952 | } |
953 | |
954 | /** |
955 | * atl1c_free_ring_resources - Free Tx / RX descriptor Resources |
956 | * @adapter: board private structure |
957 | * |
958 | * Free all transmit software resources |
959 | */ |
960 | static void atl1c_free_ring_resources(struct atl1c_adapter *adapter) |
961 | { |
962 | struct pci_dev *pdev = adapter->pdev; |
963 | |
964 | dma_free_coherent(dev: &pdev->dev, size: adapter->ring_header.size, |
965 | cpu_addr: adapter->ring_header.desc, dma_handle: adapter->ring_header.dma); |
966 | adapter->ring_header.desc = NULL; |
967 | |
968 | /* Note: just free tdp_ring.buffer_info, |
969 | * it contain rfd_ring.buffer_info, do not double free |
970 | */ |
971 | if (adapter->tpd_ring[0].buffer_info) { |
972 | kfree(objp: adapter->tpd_ring[0].buffer_info); |
973 | adapter->tpd_ring[0].buffer_info = NULL; |
974 | } |
975 | } |
976 | |
977 | /** |
978 | * atl1c_setup_ring_resources - allocate Tx / RX descriptor resources |
979 | * @adapter: board private structure |
980 | * |
981 | * Return 0 on success, negative on failure |
982 | */ |
983 | static int atl1c_setup_ring_resources(struct atl1c_adapter *adapter) |
984 | { |
985 | struct pci_dev *pdev = adapter->pdev; |
986 | struct atl1c_tpd_ring *tpd_ring = adapter->tpd_ring; |
987 | struct atl1c_rfd_ring *rfd_ring = adapter->rfd_ring; |
988 | struct atl1c_rrd_ring *rrd_ring = adapter->rrd_ring; |
989 | struct atl1c_ring_header * = &adapter->ring_header; |
990 | int tqc = adapter->tx_queue_count; |
991 | int rqc = adapter->rx_queue_count; |
992 | int size; |
993 | int i; |
994 | int count = 0; |
995 | u32 offset = 0; |
996 | |
997 | /* Even though only one tpd queue is actually used, the "high" |
998 | * priority tpd queue also gets initialized |
999 | */ |
1000 | if (tqc == 1) |
1001 | tqc = 2; |
1002 | |
1003 | for (i = 1; i < tqc; i++) |
1004 | tpd_ring[i].count = tpd_ring[0].count; |
1005 | |
1006 | size = sizeof(struct atl1c_buffer) * (tpd_ring->count * tqc + |
1007 | rfd_ring->count * rqc); |
1008 | tpd_ring->buffer_info = kzalloc(size, GFP_KERNEL); |
1009 | if (unlikely(!tpd_ring->buffer_info)) |
1010 | goto err_nomem; |
1011 | |
1012 | for (i = 0; i < tqc; i++) { |
1013 | tpd_ring[i].adapter = adapter; |
1014 | tpd_ring[i].num = i; |
1015 | tpd_ring[i].buffer_info = (tpd_ring->buffer_info + count); |
1016 | count += tpd_ring[i].count; |
1017 | } |
1018 | |
1019 | for (i = 0; i < rqc; i++) { |
1020 | rrd_ring[i].adapter = adapter; |
1021 | rrd_ring[i].num = i; |
1022 | rrd_ring[i].count = rfd_ring[0].count; |
1023 | rfd_ring[i].count = rfd_ring[0].count; |
1024 | rfd_ring[i].buffer_info = (tpd_ring->buffer_info + count); |
1025 | count += rfd_ring->count; |
1026 | } |
1027 | |
1028 | /* |
1029 | * real ring DMA buffer |
1030 | * each ring/block may need up to 8 bytes for alignment, hence the |
1031 | * additional bytes tacked onto the end. |
1032 | */ |
1033 | ring_header->size = |
1034 | sizeof(struct atl1c_tpd_desc) * tpd_ring->count * tqc + |
1035 | sizeof(struct atl1c_rx_free_desc) * rfd_ring->count * rqc + |
1036 | sizeof(struct atl1c_recv_ret_status) * rfd_ring->count * rqc + |
1037 | 8 * 4; |
1038 | |
1039 | ring_header->desc = dma_alloc_coherent(dev: &pdev->dev, size: ring_header->size, |
1040 | dma_handle: &ring_header->dma, GFP_KERNEL); |
1041 | if (unlikely(!ring_header->desc)) { |
1042 | dev_err(&pdev->dev, "could not get memory for DMA buffer\n" ); |
1043 | goto err_nomem; |
1044 | } |
1045 | /* init TPD ring */ |
1046 | |
1047 | tpd_ring[0].dma = roundup(ring_header->dma, 8); |
1048 | offset = tpd_ring[0].dma - ring_header->dma; |
1049 | for (i = 0; i < tqc; i++) { |
1050 | tpd_ring[i].dma = ring_header->dma + offset; |
1051 | tpd_ring[i].desc = (u8 *)ring_header->desc + offset; |
1052 | tpd_ring[i].size = |
1053 | sizeof(struct atl1c_tpd_desc) * tpd_ring[i].count; |
1054 | offset += roundup(tpd_ring[i].size, 8); |
1055 | } |
1056 | for (i = 0; i < rqc; i++) { |
1057 | /* init RFD ring */ |
1058 | rfd_ring[i].dma = ring_header->dma + offset; |
1059 | rfd_ring[i].desc = (u8 *)ring_header->desc + offset; |
1060 | rfd_ring[i].size = sizeof(struct atl1c_rx_free_desc) * |
1061 | rfd_ring[i].count; |
1062 | offset += roundup(rfd_ring[i].size, 8); |
1063 | |
1064 | /* init RRD ring */ |
1065 | rrd_ring[i].dma = ring_header->dma + offset; |
1066 | rrd_ring[i].desc = (u8 *)ring_header->desc + offset; |
1067 | rrd_ring[i].size = sizeof(struct atl1c_recv_ret_status) * |
1068 | rrd_ring[i].count; |
1069 | offset += roundup(rrd_ring[i].size, 8); |
1070 | } |
1071 | |
1072 | return 0; |
1073 | |
1074 | err_nomem: |
1075 | kfree(objp: tpd_ring->buffer_info); |
1076 | return -ENOMEM; |
1077 | } |
1078 | |
1079 | static void atl1c_configure_des_ring(struct atl1c_adapter *adapter) |
1080 | { |
1081 | struct atl1c_hw *hw = &adapter->hw; |
1082 | struct atl1c_rfd_ring *rfd_ring = adapter->rfd_ring; |
1083 | struct atl1c_rrd_ring *rrd_ring = adapter->rrd_ring; |
1084 | struct atl1c_tpd_ring *tpd_ring = adapter->tpd_ring; |
1085 | int i; |
1086 | int tx_queue_count = adapter->tx_queue_count; |
1087 | |
1088 | if (tx_queue_count == 1) |
1089 | tx_queue_count = 2; |
1090 | |
1091 | /* TPD */ |
1092 | AT_WRITE_REG(hw, REG_TX_BASE_ADDR_HI, |
1093 | (u32)((tpd_ring[0].dma & AT_DMA_HI_ADDR_MASK) >> 32)); |
1094 | /* just enable normal priority TX queue */ |
1095 | for (i = 0; i < tx_queue_count; i++) { |
1096 | AT_WRITE_REG(hw, atl1c_qregs[i].tpd_addr_lo, |
1097 | (u32)(tpd_ring[i].dma & AT_DMA_LO_ADDR_MASK)); |
1098 | } |
1099 | AT_WRITE_REG(hw, REG_TPD_RING_SIZE, |
1100 | (u32)(tpd_ring[0].count & TPD_RING_SIZE_MASK)); |
1101 | |
1102 | |
1103 | /* RFD */ |
1104 | AT_WRITE_REG(hw, REG_RX_BASE_ADDR_HI, |
1105 | (u32)((rfd_ring->dma & AT_DMA_HI_ADDR_MASK) >> 32)); |
1106 | for (i = 0; i < adapter->rx_queue_count; i++) { |
1107 | AT_WRITE_REG(hw, atl1c_qregs[i].rfd_addr_lo, |
1108 | (u32)(rfd_ring[i].dma & AT_DMA_LO_ADDR_MASK)); |
1109 | } |
1110 | |
1111 | AT_WRITE_REG(hw, REG_RFD_RING_SIZE, |
1112 | rfd_ring->count & RFD_RING_SIZE_MASK); |
1113 | AT_WRITE_REG(hw, REG_RX_BUF_SIZE, |
1114 | adapter->rx_buffer_len & RX_BUF_SIZE_MASK); |
1115 | |
1116 | /* RRD */ |
1117 | for (i = 0; i < adapter->rx_queue_count; i++) { |
1118 | AT_WRITE_REG(hw, atl1c_qregs[i].rrd_addr_lo, |
1119 | (u32)(rrd_ring[i].dma & AT_DMA_LO_ADDR_MASK)); |
1120 | } |
1121 | AT_WRITE_REG(hw, REG_RRD_RING_SIZE, |
1122 | (rrd_ring->count & RRD_RING_SIZE_MASK)); |
1123 | |
1124 | if (hw->nic_type == athr_l2c_b) { |
1125 | AT_WRITE_REG(hw, REG_SRAM_RXF_LEN, 0x02a0L); |
1126 | AT_WRITE_REG(hw, REG_SRAM_TXF_LEN, 0x0100L); |
1127 | AT_WRITE_REG(hw, REG_SRAM_RXF_ADDR, 0x029f0000L); |
1128 | AT_WRITE_REG(hw, REG_SRAM_RFD0_INFO, 0x02bf02a0L); |
1129 | AT_WRITE_REG(hw, REG_SRAM_TXF_ADDR, 0x03bf02c0L); |
1130 | AT_WRITE_REG(hw, REG_SRAM_TRD_ADDR, 0x03df03c0L); |
1131 | AT_WRITE_REG(hw, REG_TXF_WATER_MARK, 0); /* TX watermark, to enter l1 state.*/ |
1132 | AT_WRITE_REG(hw, REG_RXD_DMA_CTRL, 0); /* RXD threshold.*/ |
1133 | } |
1134 | /* Load all of base address above */ |
1135 | AT_WRITE_REG(hw, REG_LOAD_PTR, 1); |
1136 | } |
1137 | |
1138 | static void atl1c_configure_tx(struct atl1c_adapter *adapter) |
1139 | { |
1140 | struct atl1c_hw *hw = &adapter->hw; |
1141 | int max_pay_load; |
1142 | u16 tx_offload_thresh; |
1143 | u32 txq_ctrl_data; |
1144 | |
1145 | tx_offload_thresh = MAX_TSO_FRAME_SIZE; |
1146 | AT_WRITE_REG(hw, REG_TX_TSO_OFFLOAD_THRESH, |
1147 | (tx_offload_thresh >> 3) & TX_TSO_OFFLOAD_THRESH_MASK); |
1148 | max_pay_load = pcie_get_readrq(dev: adapter->pdev) >> 8; |
1149 | hw->dmar_block = min_t(u32, max_pay_load, hw->dmar_block); |
1150 | /* |
1151 | * if BIOS had changed the dam-read-max-length to an invalid value, |
1152 | * restore it to default value |
1153 | */ |
1154 | if (hw->dmar_block < DEVICE_CTRL_MAXRRS_MIN) { |
1155 | pcie_set_readrq(dev: adapter->pdev, rq: 128 << DEVICE_CTRL_MAXRRS_MIN); |
1156 | hw->dmar_block = DEVICE_CTRL_MAXRRS_MIN; |
1157 | } |
1158 | txq_ctrl_data = |
1159 | hw->nic_type == athr_l2c_b || hw->nic_type == athr_l2c_b2 ? |
1160 | L2CB_TXQ_CFGV : L1C_TXQ_CFGV; |
1161 | |
1162 | AT_WRITE_REG(hw, REG_TXQ_CTRL, txq_ctrl_data); |
1163 | } |
1164 | |
1165 | static void atl1c_configure_rx(struct atl1c_adapter *adapter) |
1166 | { |
1167 | struct atl1c_hw *hw = &adapter->hw; |
1168 | u32 rxq_ctrl_data; |
1169 | |
1170 | rxq_ctrl_data = (hw->rfd_burst & RXQ_RFD_BURST_NUM_MASK) << |
1171 | RXQ_RFD_BURST_NUM_SHIFT; |
1172 | |
1173 | if (hw->ctrl_flags & ATL1C_RX_IPV6_CHKSUM) |
1174 | rxq_ctrl_data |= IPV6_CHKSUM_CTRL_EN; |
1175 | |
1176 | /* aspm for gigabit */ |
1177 | if (hw->nic_type != athr_l1d_2 && (hw->device_id & 1) != 0) |
1178 | rxq_ctrl_data = FIELD_SETX(rxq_ctrl_data, ASPM_THRUPUT_LIMIT, |
1179 | ASPM_THRUPUT_LIMIT_100M); |
1180 | |
1181 | AT_WRITE_REG(hw, REG_RXQ_CTRL, rxq_ctrl_data); |
1182 | } |
1183 | |
1184 | static void atl1c_configure_dma(struct atl1c_adapter *adapter) |
1185 | { |
1186 | struct atl1c_hw *hw = &adapter->hw; |
1187 | u32 dma_ctrl_data; |
1188 | |
1189 | dma_ctrl_data = FIELDX(DMA_CTRL_RORDER_MODE, DMA_CTRL_RORDER_MODE_OUT) | |
1190 | DMA_CTRL_RREQ_PRI_DATA | |
1191 | FIELDX(DMA_CTRL_RREQ_BLEN, hw->dmar_block) | |
1192 | FIELDX(DMA_CTRL_WDLY_CNT, DMA_CTRL_WDLY_CNT_DEF) | |
1193 | FIELDX(DMA_CTRL_RDLY_CNT, DMA_CTRL_RDLY_CNT_DEF); |
1194 | |
1195 | AT_WRITE_REG(hw, REG_DMA_CTRL, dma_ctrl_data); |
1196 | } |
1197 | |
1198 | /* |
1199 | * Stop the mac, transmit and receive units |
1200 | * hw - Struct containing variables accessed by shared code |
1201 | * return : 0 or idle status (if error) |
1202 | */ |
1203 | static int atl1c_stop_mac(struct atl1c_hw *hw) |
1204 | { |
1205 | u32 data; |
1206 | |
1207 | AT_READ_REG(hw, REG_RXQ_CTRL, &data); |
1208 | data &= ~RXQ_CTRL_EN; |
1209 | AT_WRITE_REG(hw, REG_RXQ_CTRL, data); |
1210 | |
1211 | AT_READ_REG(hw, REG_TXQ_CTRL, &data); |
1212 | data &= ~TXQ_CTRL_EN; |
1213 | AT_WRITE_REG(hw, REG_TXQ_CTRL, data); |
1214 | |
1215 | atl1c_wait_until_idle(hw, IDLE_STATUS_RXQ_BUSY | IDLE_STATUS_TXQ_BUSY); |
1216 | |
1217 | AT_READ_REG(hw, REG_MAC_CTRL, &data); |
1218 | data &= ~(MAC_CTRL_TX_EN | MAC_CTRL_RX_EN); |
1219 | AT_WRITE_REG(hw, REG_MAC_CTRL, data); |
1220 | |
1221 | return (int)atl1c_wait_until_idle(hw, |
1222 | IDLE_STATUS_TXMAC_BUSY | IDLE_STATUS_RXMAC_BUSY); |
1223 | } |
1224 | |
1225 | static void atl1c_start_mac(struct atl1c_adapter *adapter) |
1226 | { |
1227 | struct atl1c_hw *hw = &adapter->hw; |
1228 | u32 mac, txq, rxq; |
1229 | |
1230 | hw->mac_duplex = adapter->link_duplex == FULL_DUPLEX; |
1231 | hw->mac_speed = adapter->link_speed == SPEED_1000 ? |
1232 | atl1c_mac_speed_1000 : atl1c_mac_speed_10_100; |
1233 | |
1234 | AT_READ_REG(hw, REG_TXQ_CTRL, &txq); |
1235 | AT_READ_REG(hw, REG_RXQ_CTRL, &rxq); |
1236 | AT_READ_REG(hw, REG_MAC_CTRL, &mac); |
1237 | |
1238 | txq |= TXQ_CTRL_EN; |
1239 | rxq |= RXQ_CTRL_EN; |
1240 | mac |= MAC_CTRL_TX_EN | MAC_CTRL_TX_FLOW | |
1241 | MAC_CTRL_RX_EN | MAC_CTRL_RX_FLOW | |
1242 | MAC_CTRL_ADD_CRC | MAC_CTRL_PAD | |
1243 | MAC_CTRL_BC_EN | MAC_CTRL_SINGLE_PAUSE_EN | |
1244 | MAC_CTRL_HASH_ALG_CRC32; |
1245 | if (hw->mac_duplex) |
1246 | mac |= MAC_CTRL_DUPLX; |
1247 | else |
1248 | mac &= ~MAC_CTRL_DUPLX; |
1249 | mac = FIELD_SETX(mac, MAC_CTRL_SPEED, hw->mac_speed); |
1250 | mac = FIELD_SETX(mac, MAC_CTRL_PRMLEN, hw->preamble_len); |
1251 | |
1252 | AT_WRITE_REG(hw, REG_TXQ_CTRL, txq); |
1253 | AT_WRITE_REG(hw, REG_RXQ_CTRL, rxq); |
1254 | AT_WRITE_REG(hw, REG_MAC_CTRL, mac); |
1255 | } |
1256 | |
1257 | /* |
1258 | * Reset the transmit and receive units; mask and clear all interrupts. |
1259 | * hw - Struct containing variables accessed by shared code |
1260 | * return : 0 or idle status (if error) |
1261 | */ |
1262 | static int atl1c_reset_mac(struct atl1c_hw *hw) |
1263 | { |
1264 | struct atl1c_adapter *adapter = hw->adapter; |
1265 | struct pci_dev *pdev = adapter->pdev; |
1266 | u32 ctrl_data = 0; |
1267 | |
1268 | atl1c_stop_mac(hw); |
1269 | /* |
1270 | * Issue Soft Reset to the MAC. This will reset the chip's |
1271 | * transmit, receive, DMA. It will not effect |
1272 | * the current PCI configuration. The global reset bit is self- |
1273 | * clearing, and should clear within a microsecond. |
1274 | */ |
1275 | AT_READ_REG(hw, REG_MASTER_CTRL, &ctrl_data); |
1276 | ctrl_data |= MASTER_CTRL_OOB_DIS; |
1277 | AT_WRITE_REG(hw, REG_MASTER_CTRL, ctrl_data | MASTER_CTRL_SOFT_RST); |
1278 | |
1279 | AT_WRITE_FLUSH(hw); |
1280 | msleep(msecs: 10); |
1281 | /* Wait at least 10ms for All module to be Idle */ |
1282 | |
1283 | if (atl1c_wait_until_idle(hw, IDLE_STATUS_MASK)) { |
1284 | dev_err(&pdev->dev, |
1285 | "MAC state machine can't be idle since" |
1286 | " disabled for 10ms second\n" ); |
1287 | return -1; |
1288 | } |
1289 | AT_WRITE_REG(hw, REG_MASTER_CTRL, ctrl_data); |
1290 | |
1291 | /* driver control speed/duplex */ |
1292 | AT_READ_REG(hw, REG_MAC_CTRL, &ctrl_data); |
1293 | AT_WRITE_REG(hw, REG_MAC_CTRL, ctrl_data | MAC_CTRL_SPEED_MODE_SW); |
1294 | |
1295 | /* clk switch setting */ |
1296 | AT_READ_REG(hw, REG_SERDES, &ctrl_data); |
1297 | switch (hw->nic_type) { |
1298 | case athr_l2c_b: |
1299 | ctrl_data &= ~(SERDES_PHY_CLK_SLOWDOWN | |
1300 | SERDES_MAC_CLK_SLOWDOWN); |
1301 | AT_WRITE_REG(hw, REG_SERDES, ctrl_data); |
1302 | break; |
1303 | case athr_l2c_b2: |
1304 | case athr_l1d_2: |
1305 | ctrl_data |= SERDES_PHY_CLK_SLOWDOWN | SERDES_MAC_CLK_SLOWDOWN; |
1306 | AT_WRITE_REG(hw, REG_SERDES, ctrl_data); |
1307 | break; |
1308 | default: |
1309 | break; |
1310 | } |
1311 | |
1312 | return 0; |
1313 | } |
1314 | |
1315 | static void atl1c_disable_l0s_l1(struct atl1c_hw *hw) |
1316 | { |
1317 | u16 ctrl_flags = hw->ctrl_flags; |
1318 | |
1319 | hw->ctrl_flags &= ~(ATL1C_ASPM_L0S_SUPPORT | ATL1C_ASPM_L1_SUPPORT); |
1320 | atl1c_set_aspm(hw, SPEED_0); |
1321 | hw->ctrl_flags = ctrl_flags; |
1322 | } |
1323 | |
1324 | /* |
1325 | * Set ASPM state. |
1326 | * Enable/disable L0s/L1 depend on link state. |
1327 | */ |
1328 | static void atl1c_set_aspm(struct atl1c_hw *hw, u16 link_speed) |
1329 | { |
1330 | u32 pm_ctrl_data; |
1331 | u32 link_l1_timer; |
1332 | |
1333 | AT_READ_REG(hw, REG_PM_CTRL, &pm_ctrl_data); |
1334 | pm_ctrl_data &= ~(PM_CTRL_ASPM_L1_EN | |
1335 | PM_CTRL_ASPM_L0S_EN | |
1336 | PM_CTRL_MAC_ASPM_CHK); |
1337 | /* L1 timer */ |
1338 | if (hw->nic_type == athr_l2c_b2 || hw->nic_type == athr_l1d_2) { |
1339 | pm_ctrl_data &= ~PMCTRL_TXL1_AFTER_L0S; |
1340 | link_l1_timer = |
1341 | link_speed == SPEED_1000 || link_speed == SPEED_100 ? |
1342 | L1D_PMCTRL_L1_ENTRY_TM_16US : 1; |
1343 | pm_ctrl_data = FIELD_SETX(pm_ctrl_data, |
1344 | L1D_PMCTRL_L1_ENTRY_TM, link_l1_timer); |
1345 | } else { |
1346 | link_l1_timer = hw->nic_type == athr_l2c_b ? |
1347 | L2CB1_PM_CTRL_L1_ENTRY_TM : L1C_PM_CTRL_L1_ENTRY_TM; |
1348 | if (link_speed != SPEED_1000 && link_speed != SPEED_100) |
1349 | link_l1_timer = 1; |
1350 | pm_ctrl_data = FIELD_SETX(pm_ctrl_data, |
1351 | PM_CTRL_L1_ENTRY_TIMER, link_l1_timer); |
1352 | } |
1353 | |
1354 | /* L0S/L1 enable */ |
1355 | if ((hw->ctrl_flags & ATL1C_ASPM_L0S_SUPPORT) && link_speed != SPEED_0) |
1356 | pm_ctrl_data |= PM_CTRL_ASPM_L0S_EN | PM_CTRL_MAC_ASPM_CHK; |
1357 | if (hw->ctrl_flags & ATL1C_ASPM_L1_SUPPORT) |
1358 | pm_ctrl_data |= PM_CTRL_ASPM_L1_EN | PM_CTRL_MAC_ASPM_CHK; |
1359 | |
1360 | /* l2cb & l1d & l2cb2 & l1d2 */ |
1361 | if (hw->nic_type == athr_l2c_b || hw->nic_type == athr_l1d || |
1362 | hw->nic_type == athr_l2c_b2 || hw->nic_type == athr_l1d_2) { |
1363 | pm_ctrl_data = FIELD_SETX(pm_ctrl_data, |
1364 | PM_CTRL_PM_REQ_TIMER, PM_CTRL_PM_REQ_TO_DEF); |
1365 | pm_ctrl_data |= PM_CTRL_RCVR_WT_TIMER | |
1366 | PM_CTRL_SERDES_PD_EX_L1 | |
1367 | PM_CTRL_CLK_SWH_L1; |
1368 | pm_ctrl_data &= ~(PM_CTRL_SERDES_L1_EN | |
1369 | PM_CTRL_SERDES_PLL_L1_EN | |
1370 | PM_CTRL_SERDES_BUFS_RX_L1_EN | |
1371 | PM_CTRL_SA_DLY_EN | |
1372 | PM_CTRL_HOTRST); |
1373 | /* disable l0s if link down or l2cb */ |
1374 | if (link_speed == SPEED_0 || hw->nic_type == athr_l2c_b) |
1375 | pm_ctrl_data &= ~PM_CTRL_ASPM_L0S_EN; |
1376 | } else { /* l1c */ |
1377 | pm_ctrl_data = |
1378 | FIELD_SETX(pm_ctrl_data, PM_CTRL_L1_ENTRY_TIMER, 0); |
1379 | if (link_speed != SPEED_0) { |
1380 | pm_ctrl_data |= PM_CTRL_SERDES_L1_EN | |
1381 | PM_CTRL_SERDES_PLL_L1_EN | |
1382 | PM_CTRL_SERDES_BUFS_RX_L1_EN; |
1383 | pm_ctrl_data &= ~(PM_CTRL_SERDES_PD_EX_L1 | |
1384 | PM_CTRL_CLK_SWH_L1 | |
1385 | PM_CTRL_ASPM_L0S_EN | |
1386 | PM_CTRL_ASPM_L1_EN); |
1387 | } else { /* link down */ |
1388 | pm_ctrl_data |= PM_CTRL_CLK_SWH_L1; |
1389 | pm_ctrl_data &= ~(PM_CTRL_SERDES_L1_EN | |
1390 | PM_CTRL_SERDES_PLL_L1_EN | |
1391 | PM_CTRL_SERDES_BUFS_RX_L1_EN | |
1392 | PM_CTRL_ASPM_L0S_EN); |
1393 | } |
1394 | } |
1395 | AT_WRITE_REG(hw, REG_PM_CTRL, pm_ctrl_data); |
1396 | |
1397 | return; |
1398 | } |
1399 | |
1400 | /** |
1401 | * atl1c_configure_mac - Configure Transmit&Receive Unit after Reset |
1402 | * @adapter: board private structure |
1403 | * |
1404 | * Configure the Tx /Rx unit of the MAC after a reset. |
1405 | */ |
1406 | static int atl1c_configure_mac(struct atl1c_adapter *adapter) |
1407 | { |
1408 | struct atl1c_hw *hw = &adapter->hw; |
1409 | u32 master_ctrl_data = 0; |
1410 | u32 intr_modrt_data; |
1411 | u32 data; |
1412 | |
1413 | AT_READ_REG(hw, REG_MASTER_CTRL, &master_ctrl_data); |
1414 | master_ctrl_data &= ~(MASTER_CTRL_TX_ITIMER_EN | |
1415 | MASTER_CTRL_RX_ITIMER_EN | |
1416 | MASTER_CTRL_INT_RDCLR); |
1417 | /* clear interrupt status */ |
1418 | AT_WRITE_REG(hw, REG_ISR, 0xFFFFFFFF); |
1419 | /* Clear any WOL status */ |
1420 | AT_WRITE_REG(hw, REG_WOL_CTRL, 0); |
1421 | /* set Interrupt Clear Timer |
1422 | * HW will enable self to assert interrupt event to system after |
1423 | * waiting x-time for software to notify it accept interrupt. |
1424 | */ |
1425 | |
1426 | data = CLK_GATING_EN_ALL; |
1427 | if (hw->ctrl_flags & ATL1C_CLK_GATING_EN) { |
1428 | if (hw->nic_type == athr_l2c_b) |
1429 | data &= ~CLK_GATING_RXMAC_EN; |
1430 | } else |
1431 | data = 0; |
1432 | AT_WRITE_REG(hw, REG_CLK_GATING_CTRL, data); |
1433 | |
1434 | AT_WRITE_REG(hw, REG_INT_RETRIG_TIMER, |
1435 | hw->ict & INT_RETRIG_TIMER_MASK); |
1436 | |
1437 | atl1c_configure_des_ring(adapter); |
1438 | |
1439 | if (hw->ctrl_flags & ATL1C_INTR_MODRT_ENABLE) { |
1440 | intr_modrt_data = (hw->tx_imt & IRQ_MODRT_TIMER_MASK) << |
1441 | IRQ_MODRT_TX_TIMER_SHIFT; |
1442 | intr_modrt_data |= (hw->rx_imt & IRQ_MODRT_TIMER_MASK) << |
1443 | IRQ_MODRT_RX_TIMER_SHIFT; |
1444 | AT_WRITE_REG(hw, REG_IRQ_MODRT_TIMER_INIT, intr_modrt_data); |
1445 | master_ctrl_data |= |
1446 | MASTER_CTRL_TX_ITIMER_EN | MASTER_CTRL_RX_ITIMER_EN; |
1447 | } |
1448 | |
1449 | if (hw->ctrl_flags & ATL1C_INTR_CLEAR_ON_READ) |
1450 | master_ctrl_data |= MASTER_CTRL_INT_RDCLR; |
1451 | |
1452 | master_ctrl_data |= MASTER_CTRL_SA_TIMER_EN; |
1453 | AT_WRITE_REG(hw, REG_MASTER_CTRL, master_ctrl_data); |
1454 | |
1455 | AT_WRITE_REG(hw, REG_SMB_STAT_TIMER, |
1456 | hw->smb_timer & SMB_STAT_TIMER_MASK); |
1457 | |
1458 | /* set MTU */ |
1459 | AT_WRITE_REG(hw, REG_MTU, hw->max_frame_size + ETH_HLEN + |
1460 | VLAN_HLEN + ETH_FCS_LEN); |
1461 | |
1462 | atl1c_configure_tx(adapter); |
1463 | atl1c_configure_rx(adapter); |
1464 | atl1c_configure_dma(adapter); |
1465 | |
1466 | return 0; |
1467 | } |
1468 | |
1469 | static int atl1c_configure(struct atl1c_adapter *adapter) |
1470 | { |
1471 | struct net_device *netdev = adapter->netdev; |
1472 | int num; |
1473 | int i; |
1474 | |
1475 | if (adapter->hw.nic_type == athr_mt) { |
1476 | u32 mode; |
1477 | |
1478 | AT_READ_REG(&adapter->hw, REG_MT_MODE, &mode); |
1479 | if (adapter->rx_queue_count == 4) |
1480 | mode |= MT_MODE_4Q; |
1481 | else |
1482 | mode &= ~MT_MODE_4Q; |
1483 | AT_WRITE_REG(&adapter->hw, REG_MT_MODE, mode); |
1484 | } |
1485 | |
1486 | atl1c_init_ring_ptrs(adapter); |
1487 | atl1c_set_multi(netdev); |
1488 | atl1c_restore_vlan(adapter); |
1489 | |
1490 | for (i = 0; i < adapter->rx_queue_count; ++i) { |
1491 | num = atl1c_alloc_rx_buffer(adapter, queue: i, napi_mode: false); |
1492 | if (unlikely(num == 0)) |
1493 | return -ENOMEM; |
1494 | } |
1495 | |
1496 | if (atl1c_configure_mac(adapter)) |
1497 | return -EIO; |
1498 | |
1499 | return 0; |
1500 | } |
1501 | |
1502 | static void atl1c_update_hw_stats(struct atl1c_adapter *adapter) |
1503 | { |
1504 | u16 hw_reg_addr = 0; |
1505 | unsigned long *stats_item = NULL; |
1506 | u32 data; |
1507 | |
1508 | /* update rx status */ |
1509 | hw_reg_addr = REG_MAC_RX_STATUS_BIN; |
1510 | stats_item = &adapter->hw_stats.rx_ok; |
1511 | while (hw_reg_addr <= REG_MAC_RX_STATUS_END) { |
1512 | AT_READ_REG(&adapter->hw, hw_reg_addr, &data); |
1513 | *stats_item += data; |
1514 | stats_item++; |
1515 | hw_reg_addr += 4; |
1516 | } |
1517 | /* update tx status */ |
1518 | hw_reg_addr = REG_MAC_TX_STATUS_BIN; |
1519 | stats_item = &adapter->hw_stats.tx_ok; |
1520 | while (hw_reg_addr <= REG_MAC_TX_STATUS_END) { |
1521 | AT_READ_REG(&adapter->hw, hw_reg_addr, &data); |
1522 | *stats_item += data; |
1523 | stats_item++; |
1524 | hw_reg_addr += 4; |
1525 | } |
1526 | } |
1527 | |
1528 | /** |
1529 | * atl1c_get_stats - Get System Network Statistics |
1530 | * @netdev: network interface device structure |
1531 | * |
1532 | * Returns the address of the device statistics structure. |
1533 | * The statistics are actually updated from the timer callback. |
1534 | */ |
1535 | static struct net_device_stats *atl1c_get_stats(struct net_device *netdev) |
1536 | { |
1537 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
1538 | struct atl1c_hw_stats *hw_stats = &adapter->hw_stats; |
1539 | struct net_device_stats *net_stats = &netdev->stats; |
1540 | |
1541 | atl1c_update_hw_stats(adapter); |
1542 | net_stats->rx_bytes = hw_stats->rx_byte_cnt; |
1543 | net_stats->tx_bytes = hw_stats->tx_byte_cnt; |
1544 | net_stats->multicast = hw_stats->rx_mcast; |
1545 | net_stats->collisions = hw_stats->tx_1_col + |
1546 | hw_stats->tx_2_col + |
1547 | hw_stats->tx_late_col + |
1548 | hw_stats->tx_abort_col; |
1549 | |
1550 | net_stats->rx_errors = hw_stats->rx_frag + |
1551 | hw_stats->rx_fcs_err + |
1552 | hw_stats->rx_len_err + |
1553 | hw_stats->rx_sz_ov + |
1554 | hw_stats->rx_rrd_ov + |
1555 | hw_stats->rx_align_err + |
1556 | hw_stats->rx_rxf_ov; |
1557 | |
1558 | net_stats->rx_fifo_errors = hw_stats->rx_rxf_ov; |
1559 | net_stats->rx_length_errors = hw_stats->rx_len_err; |
1560 | net_stats->rx_crc_errors = hw_stats->rx_fcs_err; |
1561 | net_stats->rx_frame_errors = hw_stats->rx_align_err; |
1562 | net_stats->rx_dropped = hw_stats->rx_rrd_ov; |
1563 | |
1564 | net_stats->tx_errors = hw_stats->tx_late_col + |
1565 | hw_stats->tx_abort_col + |
1566 | hw_stats->tx_underrun + |
1567 | hw_stats->tx_trunc; |
1568 | |
1569 | net_stats->tx_fifo_errors = hw_stats->tx_underrun; |
1570 | net_stats->tx_aborted_errors = hw_stats->tx_abort_col; |
1571 | net_stats->tx_window_errors = hw_stats->tx_late_col; |
1572 | |
1573 | net_stats->rx_packets = hw_stats->rx_ok + net_stats->rx_errors; |
1574 | net_stats->tx_packets = hw_stats->tx_ok + net_stats->tx_errors; |
1575 | |
1576 | return net_stats; |
1577 | } |
1578 | |
1579 | static inline void atl1c_clear_phy_int(struct atl1c_adapter *adapter) |
1580 | { |
1581 | u16 phy_data; |
1582 | |
1583 | spin_lock(lock: &adapter->mdio_lock); |
1584 | atl1c_read_phy_reg(hw: &adapter->hw, MII_ISR, phy_data: &phy_data); |
1585 | spin_unlock(lock: &adapter->mdio_lock); |
1586 | } |
1587 | |
1588 | static int atl1c_clean_tx(struct napi_struct *napi, int budget) |
1589 | { |
1590 | struct atl1c_tpd_ring *tpd_ring = |
1591 | container_of(napi, struct atl1c_tpd_ring, napi); |
1592 | struct atl1c_adapter *adapter = tpd_ring->adapter; |
1593 | struct netdev_queue *txq = |
1594 | netdev_get_tx_queue(dev: napi->dev, index: tpd_ring->num); |
1595 | struct atl1c_buffer *buffer_info; |
1596 | struct pci_dev *pdev = adapter->pdev; |
1597 | u16 next_to_clean = atomic_read(v: &tpd_ring->next_to_clean); |
1598 | u16 hw_next_to_clean; |
1599 | unsigned int total_bytes = 0, total_packets = 0; |
1600 | unsigned long flags; |
1601 | |
1602 | AT_READ_REGW(&adapter->hw, atl1c_qregs[tpd_ring->num].tpd_cons, |
1603 | &hw_next_to_clean); |
1604 | |
1605 | while (next_to_clean != hw_next_to_clean) { |
1606 | buffer_info = &tpd_ring->buffer_info[next_to_clean]; |
1607 | if (buffer_info->skb) { |
1608 | total_bytes += buffer_info->skb->len; |
1609 | total_packets++; |
1610 | } |
1611 | atl1c_clean_buffer(pdev, buffer_info, budget); |
1612 | if (++next_to_clean == tpd_ring->count) |
1613 | next_to_clean = 0; |
1614 | atomic_set(v: &tpd_ring->next_to_clean, i: next_to_clean); |
1615 | } |
1616 | |
1617 | netdev_tx_completed_queue(dev_queue: txq, pkts: total_packets, bytes: total_bytes); |
1618 | |
1619 | if (netif_tx_queue_stopped(dev_queue: txq) && netif_carrier_ok(dev: adapter->netdev)) |
1620 | netif_tx_wake_queue(dev_queue: txq); |
1621 | |
1622 | if (total_packets < budget) { |
1623 | napi_complete_done(n: napi, work_done: total_packets); |
1624 | spin_lock_irqsave(&adapter->hw.intr_mask_lock, flags); |
1625 | adapter->hw.intr_mask |= atl1c_qregs[tpd_ring->num].tx_isr; |
1626 | AT_WRITE_REG(&adapter->hw, REG_IMR, adapter->hw.intr_mask); |
1627 | spin_unlock_irqrestore(lock: &adapter->hw.intr_mask_lock, flags); |
1628 | return total_packets; |
1629 | } |
1630 | return budget; |
1631 | } |
1632 | |
1633 | static void atl1c_intr_rx_tx(struct atl1c_adapter *adapter, u32 status) |
1634 | { |
1635 | struct atl1c_hw *hw = &adapter->hw; |
1636 | u32 intr_mask; |
1637 | int i; |
1638 | |
1639 | spin_lock(lock: &hw->intr_mask_lock); |
1640 | intr_mask = hw->intr_mask; |
1641 | for (i = 0; i < adapter->rx_queue_count; ++i) { |
1642 | if (!(status & atl1c_qregs[i].rx_isr)) |
1643 | continue; |
1644 | if (napi_schedule_prep(n: &adapter->rrd_ring[i].napi)) { |
1645 | intr_mask &= ~atl1c_qregs[i].rx_isr; |
1646 | __napi_schedule(n: &adapter->rrd_ring[i].napi); |
1647 | } |
1648 | } |
1649 | for (i = 0; i < adapter->tx_queue_count; ++i) { |
1650 | if (!(status & atl1c_qregs[i].tx_isr)) |
1651 | continue; |
1652 | if (napi_schedule_prep(n: &adapter->tpd_ring[i].napi)) { |
1653 | intr_mask &= ~atl1c_qregs[i].tx_isr; |
1654 | __napi_schedule(n: &adapter->tpd_ring[i].napi); |
1655 | } |
1656 | } |
1657 | |
1658 | if (hw->intr_mask != intr_mask) { |
1659 | hw->intr_mask = intr_mask; |
1660 | AT_WRITE_REG(hw, REG_IMR, hw->intr_mask); |
1661 | } |
1662 | spin_unlock(lock: &hw->intr_mask_lock); |
1663 | } |
1664 | |
1665 | /** |
1666 | * atl1c_intr - Interrupt Handler |
1667 | * @irq: interrupt number |
1668 | * @data: pointer to a network interface device structure |
1669 | */ |
1670 | static irqreturn_t atl1c_intr(int irq, void *data) |
1671 | { |
1672 | struct net_device *netdev = data; |
1673 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
1674 | struct pci_dev *pdev = adapter->pdev; |
1675 | struct atl1c_hw *hw = &adapter->hw; |
1676 | int max_ints = AT_MAX_INT_WORK; |
1677 | int handled = IRQ_NONE; |
1678 | u32 status; |
1679 | u32 reg_data; |
1680 | |
1681 | do { |
1682 | AT_READ_REG(hw, REG_ISR, ®_data); |
1683 | status = reg_data & hw->intr_mask; |
1684 | |
1685 | if (status == 0 || (status & ISR_DIS_INT) != 0) { |
1686 | if (max_ints != AT_MAX_INT_WORK) |
1687 | handled = IRQ_HANDLED; |
1688 | break; |
1689 | } |
1690 | /* link event */ |
1691 | if (status & ISR_GPHY) |
1692 | atl1c_clear_phy_int(adapter); |
1693 | /* Ack ISR */ |
1694 | AT_WRITE_REG(hw, REG_ISR, status | ISR_DIS_INT); |
1695 | if (status & (ISR_RX_PKT | ISR_TX_PKT)) |
1696 | atl1c_intr_rx_tx(adapter, status); |
1697 | |
1698 | handled = IRQ_HANDLED; |
1699 | /* check if PCIE PHY Link down */ |
1700 | if (status & ISR_ERROR) { |
1701 | if (netif_msg_hw(adapter)) |
1702 | dev_err(&pdev->dev, |
1703 | "atl1c hardware error (status = 0x%x)\n" , |
1704 | status & ISR_ERROR); |
1705 | /* reset MAC */ |
1706 | set_bit(ATL1C_WORK_EVENT_RESET, addr: &adapter->work_event); |
1707 | schedule_work(work: &adapter->common_task); |
1708 | return IRQ_HANDLED; |
1709 | } |
1710 | |
1711 | if (status & ISR_OVER) |
1712 | if (netif_msg_intr(adapter)) |
1713 | dev_warn(&pdev->dev, |
1714 | "TX/RX overflow (status = 0x%x)\n" , |
1715 | status & ISR_OVER); |
1716 | |
1717 | /* link event */ |
1718 | if (status & (ISR_GPHY | ISR_MANUAL)) { |
1719 | netdev->stats.tx_carrier_errors++; |
1720 | atl1c_link_chg_event(adapter); |
1721 | break; |
1722 | } |
1723 | |
1724 | } while (--max_ints > 0); |
1725 | /* re-enable Interrupt*/ |
1726 | AT_WRITE_REG(&adapter->hw, REG_ISR, 0); |
1727 | return handled; |
1728 | } |
1729 | |
1730 | static inline void atl1c_rx_checksum(struct atl1c_adapter *adapter, |
1731 | struct sk_buff *skb, struct atl1c_recv_ret_status *prrs) |
1732 | { |
1733 | if (adapter->hw.nic_type == athr_mt) { |
1734 | if (prrs->word3 & RRS_MT_PROT_ID_TCPUDP) |
1735 | skb->ip_summed = CHECKSUM_UNNECESSARY; |
1736 | return; |
1737 | } |
1738 | /* |
1739 | * The pid field in RRS in not correct sometimes, so we |
1740 | * cannot figure out if the packet is fragmented or not, |
1741 | * so we tell the KERNEL CHECKSUM_NONE |
1742 | */ |
1743 | skb_checksum_none_assert(skb); |
1744 | } |
1745 | |
1746 | static int atl1c_alloc_rx_buffer(struct atl1c_adapter *adapter, u32 queue, |
1747 | bool napi_mode) |
1748 | { |
1749 | struct atl1c_rfd_ring *rfd_ring = &adapter->rfd_ring[queue]; |
1750 | struct atl1c_rrd_ring *rrd_ring = &adapter->rrd_ring[queue]; |
1751 | struct pci_dev *pdev = adapter->pdev; |
1752 | struct atl1c_buffer *buffer_info, *next_info; |
1753 | struct sk_buff *skb; |
1754 | void *vir_addr = NULL; |
1755 | u16 num_alloc = 0; |
1756 | u16 rfd_next_to_use, next_next; |
1757 | struct atl1c_rx_free_desc *rfd_desc; |
1758 | dma_addr_t mapping; |
1759 | |
1760 | next_next = rfd_next_to_use = rfd_ring->next_to_use; |
1761 | if (++next_next == rfd_ring->count) |
1762 | next_next = 0; |
1763 | buffer_info = &rfd_ring->buffer_info[rfd_next_to_use]; |
1764 | next_info = &rfd_ring->buffer_info[next_next]; |
1765 | |
1766 | while (next_info->flags & ATL1C_BUFFER_FREE) { |
1767 | rfd_desc = ATL1C_RFD_DESC(rfd_ring, rfd_next_to_use); |
1768 | |
1769 | /* When DMA RX address is set to something like |
1770 | * 0x....fc0, it will be very likely to cause DMA |
1771 | * RFD overflow issue. |
1772 | * |
1773 | * To work around it, we apply rx skb with 64 bytes |
1774 | * longer space, and offset the address whenever |
1775 | * 0x....fc0 is detected. |
1776 | */ |
1777 | if (likely(napi_mode)) |
1778 | skb = napi_alloc_skb(napi: &rrd_ring->napi, length: adapter->rx_buffer_len + 64); |
1779 | else |
1780 | skb = netdev_alloc_skb(dev: adapter->netdev, length: adapter->rx_buffer_len + 64); |
1781 | if (unlikely(!skb)) { |
1782 | if (netif_msg_rx_err(adapter)) |
1783 | dev_warn(&pdev->dev, "alloc rx buffer failed\n" ); |
1784 | break; |
1785 | } |
1786 | |
1787 | if (((unsigned long)skb->data & 0xfff) == 0xfc0) |
1788 | skb_reserve(skb, len: 64); |
1789 | |
1790 | /* |
1791 | * Make buffer alignment 2 beyond a 16 byte boundary |
1792 | * this will result in a 16 byte aligned IP header after |
1793 | * the 14 byte MAC header is removed |
1794 | */ |
1795 | vir_addr = skb->data; |
1796 | ATL1C_SET_BUFFER_STATE(buffer_info, ATL1C_BUFFER_BUSY); |
1797 | buffer_info->skb = skb; |
1798 | buffer_info->length = adapter->rx_buffer_len; |
1799 | mapping = dma_map_single(&pdev->dev, vir_addr, |
1800 | buffer_info->length, DMA_FROM_DEVICE); |
1801 | if (unlikely(dma_mapping_error(&pdev->dev, mapping))) { |
1802 | dev_kfree_skb(skb); |
1803 | buffer_info->skb = NULL; |
1804 | buffer_info->length = 0; |
1805 | ATL1C_SET_BUFFER_STATE(buffer_info, ATL1C_BUFFER_FREE); |
1806 | netif_warn(adapter, rx_err, adapter->netdev, "RX dma_map_single failed" ); |
1807 | break; |
1808 | } |
1809 | buffer_info->dma = mapping; |
1810 | ATL1C_SET_PCIMAP_TYPE(buffer_info, ATL1C_PCIMAP_SINGLE, |
1811 | ATL1C_PCIMAP_FROMDEVICE); |
1812 | rfd_desc->buffer_addr = cpu_to_le64(buffer_info->dma); |
1813 | rfd_next_to_use = next_next; |
1814 | if (++next_next == rfd_ring->count) |
1815 | next_next = 0; |
1816 | buffer_info = &rfd_ring->buffer_info[rfd_next_to_use]; |
1817 | next_info = &rfd_ring->buffer_info[next_next]; |
1818 | num_alloc++; |
1819 | } |
1820 | |
1821 | if (num_alloc) { |
1822 | /* TODO: update mailbox here */ |
1823 | wmb(); |
1824 | rfd_ring->next_to_use = rfd_next_to_use; |
1825 | AT_WRITE_REG(&adapter->hw, atl1c_qregs[queue].rfd_prod, |
1826 | rfd_ring->next_to_use & MB_RFDX_PROD_IDX_MASK); |
1827 | } |
1828 | |
1829 | return num_alloc; |
1830 | } |
1831 | |
1832 | static void atl1c_clean_rrd(struct atl1c_rrd_ring *rrd_ring, |
1833 | struct atl1c_recv_ret_status *rrs, u16 num) |
1834 | { |
1835 | u16 i; |
1836 | /* the relationship between rrd and rfd is one map one */ |
1837 | for (i = 0; i < num; i++, rrs = ATL1C_RRD_DESC(rrd_ring, |
1838 | rrd_ring->next_to_clean)) { |
1839 | rrs->word3 &= ~RRS_RXD_UPDATED; |
1840 | if (++rrd_ring->next_to_clean == rrd_ring->count) |
1841 | rrd_ring->next_to_clean = 0; |
1842 | } |
1843 | } |
1844 | |
1845 | static void atl1c_clean_rfd(struct atl1c_rfd_ring *rfd_ring, |
1846 | struct atl1c_recv_ret_status *rrs, u16 num) |
1847 | { |
1848 | u16 i; |
1849 | u16 rfd_index; |
1850 | struct atl1c_buffer *buffer_info = rfd_ring->buffer_info; |
1851 | |
1852 | rfd_index = (rrs->word0 >> RRS_RX_RFD_INDEX_SHIFT) & |
1853 | RRS_RX_RFD_INDEX_MASK; |
1854 | for (i = 0; i < num; i++) { |
1855 | buffer_info[rfd_index].skb = NULL; |
1856 | ATL1C_SET_BUFFER_STATE(&buffer_info[rfd_index], |
1857 | ATL1C_BUFFER_FREE); |
1858 | if (++rfd_index == rfd_ring->count) |
1859 | rfd_index = 0; |
1860 | } |
1861 | rfd_ring->next_to_clean = rfd_index; |
1862 | } |
1863 | |
1864 | /** |
1865 | * atl1c_clean_rx - NAPI Rx polling callback |
1866 | * @napi: napi info |
1867 | * @budget: limit of packets to clean |
1868 | */ |
1869 | static int atl1c_clean_rx(struct napi_struct *napi, int budget) |
1870 | { |
1871 | struct atl1c_rrd_ring *rrd_ring = |
1872 | container_of(napi, struct atl1c_rrd_ring, napi); |
1873 | struct atl1c_adapter *adapter = rrd_ring->adapter; |
1874 | u16 rfd_num, rfd_index; |
1875 | u16 length; |
1876 | struct pci_dev *pdev = adapter->pdev; |
1877 | struct net_device *netdev = adapter->netdev; |
1878 | struct atl1c_rfd_ring *rfd_ring = &adapter->rfd_ring[rrd_ring->num]; |
1879 | struct sk_buff *skb; |
1880 | struct atl1c_recv_ret_status *rrs; |
1881 | struct atl1c_buffer *buffer_info; |
1882 | int work_done = 0; |
1883 | unsigned long flags; |
1884 | |
1885 | /* Keep link state information with original netdev */ |
1886 | if (!netif_carrier_ok(dev: adapter->netdev)) |
1887 | goto quit_polling; |
1888 | |
1889 | while (1) { |
1890 | if (work_done >= budget) |
1891 | break; |
1892 | rrs = ATL1C_RRD_DESC(rrd_ring, rrd_ring->next_to_clean); |
1893 | if (likely(RRS_RXD_IS_VALID(rrs->word3))) { |
1894 | rfd_num = (rrs->word0 >> RRS_RX_RFD_CNT_SHIFT) & |
1895 | RRS_RX_RFD_CNT_MASK; |
1896 | if (unlikely(rfd_num != 1)) |
1897 | /* TODO support mul rfd*/ |
1898 | if (netif_msg_rx_err(adapter)) |
1899 | dev_warn(&pdev->dev, |
1900 | "Multi rfd not support yet!\n" ); |
1901 | goto rrs_checked; |
1902 | } else { |
1903 | break; |
1904 | } |
1905 | rrs_checked: |
1906 | atl1c_clean_rrd(rrd_ring, rrs, num: rfd_num); |
1907 | if (rrs->word3 & (RRS_RX_ERR_SUM | RRS_802_3_LEN_ERR)) { |
1908 | atl1c_clean_rfd(rfd_ring, rrs, num: rfd_num); |
1909 | if (netif_msg_rx_err(adapter)) |
1910 | dev_warn(&pdev->dev, |
1911 | "wrong packet! rrs word3 is %x\n" , |
1912 | rrs->word3); |
1913 | continue; |
1914 | } |
1915 | |
1916 | length = le16_to_cpu((rrs->word3 >> RRS_PKT_SIZE_SHIFT) & |
1917 | RRS_PKT_SIZE_MASK); |
1918 | /* Good Receive */ |
1919 | if (likely(rfd_num == 1)) { |
1920 | rfd_index = (rrs->word0 >> RRS_RX_RFD_INDEX_SHIFT) & |
1921 | RRS_RX_RFD_INDEX_MASK; |
1922 | buffer_info = &rfd_ring->buffer_info[rfd_index]; |
1923 | dma_unmap_single(&pdev->dev, buffer_info->dma, |
1924 | buffer_info->length, DMA_FROM_DEVICE); |
1925 | skb = buffer_info->skb; |
1926 | } else { |
1927 | /* TODO */ |
1928 | if (netif_msg_rx_err(adapter)) |
1929 | dev_warn(&pdev->dev, |
1930 | "Multi rfd not support yet!\n" ); |
1931 | break; |
1932 | } |
1933 | atl1c_clean_rfd(rfd_ring, rrs, num: rfd_num); |
1934 | skb_put(skb, len: length - ETH_FCS_LEN); |
1935 | skb->protocol = eth_type_trans(skb, dev: netdev); |
1936 | atl1c_rx_checksum(adapter, skb, prrs: rrs); |
1937 | if (rrs->word3 & RRS_VLAN_INS) { |
1938 | u16 vlan; |
1939 | |
1940 | AT_TAG_TO_VLAN(rrs->vlan_tag, vlan); |
1941 | vlan = le16_to_cpu(vlan); |
1942 | __vlan_hwaccel_put_tag(skb, htons(ETH_P_8021Q), vlan_tci: vlan); |
1943 | } |
1944 | napi_gro_receive(napi, skb); |
1945 | |
1946 | work_done++; |
1947 | } |
1948 | if (work_done) |
1949 | atl1c_alloc_rx_buffer(adapter, queue: rrd_ring->num, napi_mode: true); |
1950 | |
1951 | if (work_done < budget) { |
1952 | quit_polling: |
1953 | napi_complete_done(n: napi, work_done); |
1954 | spin_lock_irqsave(&adapter->hw.intr_mask_lock, flags); |
1955 | adapter->hw.intr_mask |= atl1c_qregs[rrd_ring->num].rx_isr; |
1956 | AT_WRITE_REG(&adapter->hw, REG_IMR, adapter->hw.intr_mask); |
1957 | spin_unlock_irqrestore(lock: &adapter->hw.intr_mask_lock, flags); |
1958 | } |
1959 | return work_done; |
1960 | } |
1961 | |
1962 | #ifdef CONFIG_NET_POLL_CONTROLLER |
1963 | |
1964 | /* |
1965 | * Polling 'interrupt' - used by things like netconsole to send skbs |
1966 | * without having to re-enable interrupts. It's not called while |
1967 | * the interrupt routine is executing. |
1968 | */ |
1969 | static void atl1c_netpoll(struct net_device *netdev) |
1970 | { |
1971 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
1972 | |
1973 | disable_irq(irq: adapter->pdev->irq); |
1974 | atl1c_intr(irq: adapter->pdev->irq, data: netdev); |
1975 | enable_irq(irq: adapter->pdev->irq); |
1976 | } |
1977 | #endif |
1978 | |
1979 | static inline u16 atl1c_tpd_avail(struct atl1c_adapter *adapter, u32 queue) |
1980 | { |
1981 | struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue]; |
1982 | u16 next_to_use = 0; |
1983 | u16 next_to_clean = 0; |
1984 | |
1985 | next_to_clean = atomic_read(v: &tpd_ring->next_to_clean); |
1986 | next_to_use = tpd_ring->next_to_use; |
1987 | |
1988 | return (u16)(next_to_clean > next_to_use) ? |
1989 | (next_to_clean - next_to_use - 1) : |
1990 | (tpd_ring->count + next_to_clean - next_to_use - 1); |
1991 | } |
1992 | |
1993 | /* |
1994 | * get next usable tpd |
1995 | * Note: should call atl1c_tdp_avail to make sure |
1996 | * there is enough tpd to use |
1997 | */ |
1998 | static struct atl1c_tpd_desc *atl1c_get_tpd(struct atl1c_adapter *adapter, |
1999 | u32 queue) |
2000 | { |
2001 | struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue]; |
2002 | struct atl1c_tpd_desc *tpd_desc; |
2003 | u16 next_to_use = 0; |
2004 | |
2005 | next_to_use = tpd_ring->next_to_use; |
2006 | if (++tpd_ring->next_to_use == tpd_ring->count) |
2007 | tpd_ring->next_to_use = 0; |
2008 | tpd_desc = ATL1C_TPD_DESC(tpd_ring, next_to_use); |
2009 | memset(tpd_desc, 0, sizeof(struct atl1c_tpd_desc)); |
2010 | return tpd_desc; |
2011 | } |
2012 | |
2013 | static struct atl1c_buffer * |
2014 | atl1c_get_tx_buffer(struct atl1c_adapter *adapter, struct atl1c_tpd_desc *tpd) |
2015 | { |
2016 | struct atl1c_tpd_ring *tpd_ring = adapter->tpd_ring; |
2017 | |
2018 | return &tpd_ring->buffer_info[tpd - |
2019 | (struct atl1c_tpd_desc *)tpd_ring->desc]; |
2020 | } |
2021 | |
2022 | /* Calculate the transmit packet descript needed*/ |
2023 | static u16 atl1c_cal_tpd_req(const struct sk_buff *skb) |
2024 | { |
2025 | u16 tpd_req; |
2026 | u16 proto_hdr_len = 0; |
2027 | |
2028 | tpd_req = skb_shinfo(skb)->nr_frags + 1; |
2029 | |
2030 | if (skb_is_gso(skb)) { |
2031 | proto_hdr_len = skb_tcp_all_headers(skb); |
2032 | if (proto_hdr_len < skb_headlen(skb)) |
2033 | tpd_req++; |
2034 | if (skb_shinfo(skb)->gso_type & SKB_GSO_TCPV6) |
2035 | tpd_req++; |
2036 | } |
2037 | return tpd_req; |
2038 | } |
2039 | |
2040 | static int atl1c_tso_csum(struct atl1c_adapter *adapter, |
2041 | struct sk_buff *skb, |
2042 | struct atl1c_tpd_desc **tpd, |
2043 | u32 queue) |
2044 | { |
2045 | struct pci_dev *pdev = adapter->pdev; |
2046 | unsigned short offload_type; |
2047 | u8 hdr_len; |
2048 | u32 real_len; |
2049 | |
2050 | if (skb_is_gso(skb)) { |
2051 | int err; |
2052 | |
2053 | err = skb_cow_head(skb, headroom: 0); |
2054 | if (err < 0) |
2055 | return err; |
2056 | |
2057 | offload_type = skb_shinfo(skb)->gso_type; |
2058 | |
2059 | if (offload_type & SKB_GSO_TCPV4) { |
2060 | real_len = (((unsigned char *)ip_hdr(skb) - skb->data) |
2061 | + ntohs(ip_hdr(skb)->tot_len)); |
2062 | |
2063 | if (real_len < skb->len) { |
2064 | err = pskb_trim(skb, len: real_len); |
2065 | if (err) |
2066 | return err; |
2067 | } |
2068 | |
2069 | hdr_len = skb_tcp_all_headers(skb); |
2070 | if (unlikely(skb->len == hdr_len)) { |
2071 | /* only xsum need */ |
2072 | if (netif_msg_tx_queued(adapter)) |
2073 | dev_warn(&pdev->dev, |
2074 | "IPV4 tso with zero data??\n" ); |
2075 | goto check_sum; |
2076 | } else { |
2077 | ip_hdr(skb)->check = 0; |
2078 | tcp_hdr(skb)->check = ~csum_tcpudp_magic( |
2079 | saddr: ip_hdr(skb)->saddr, |
2080 | daddr: ip_hdr(skb)->daddr, |
2081 | len: 0, IPPROTO_TCP, sum: 0); |
2082 | (*tpd)->word1 |= 1 << TPD_IPV4_PACKET_SHIFT; |
2083 | } |
2084 | } |
2085 | |
2086 | if (offload_type & SKB_GSO_TCPV6) { |
2087 | struct atl1c_tpd_ext_desc *etpd = |
2088 | *(struct atl1c_tpd_ext_desc **)(tpd); |
2089 | |
2090 | memset(etpd, 0, sizeof(struct atl1c_tpd_ext_desc)); |
2091 | *tpd = atl1c_get_tpd(adapter, queue); |
2092 | ipv6_hdr(skb)->payload_len = 0; |
2093 | /* check payload == 0 byte ? */ |
2094 | hdr_len = skb_tcp_all_headers(skb); |
2095 | if (unlikely(skb->len == hdr_len)) { |
2096 | /* only xsum need */ |
2097 | if (netif_msg_tx_queued(adapter)) |
2098 | dev_warn(&pdev->dev, |
2099 | "IPV6 tso with zero data??\n" ); |
2100 | goto check_sum; |
2101 | } else |
2102 | tcp_v6_gso_csum_prep(skb); |
2103 | |
2104 | etpd->word1 |= 1 << TPD_LSO_EN_SHIFT; |
2105 | etpd->word1 |= 1 << TPD_LSO_VER_SHIFT; |
2106 | etpd->pkt_len = cpu_to_le32(skb->len); |
2107 | (*tpd)->word1 |= 1 << TPD_LSO_VER_SHIFT; |
2108 | } |
2109 | |
2110 | (*tpd)->word1 |= 1 << TPD_LSO_EN_SHIFT; |
2111 | (*tpd)->word1 |= (skb_transport_offset(skb) & TPD_TCPHDR_OFFSET_MASK) << |
2112 | TPD_TCPHDR_OFFSET_SHIFT; |
2113 | (*tpd)->word1 |= (skb_shinfo(skb)->gso_size & TPD_MSS_MASK) << |
2114 | TPD_MSS_SHIFT; |
2115 | return 0; |
2116 | } |
2117 | |
2118 | check_sum: |
2119 | if (likely(skb->ip_summed == CHECKSUM_PARTIAL)) { |
2120 | u8 css, cso; |
2121 | cso = skb_checksum_start_offset(skb); |
2122 | |
2123 | if (unlikely(cso & 0x1)) { |
2124 | if (netif_msg_tx_err(adapter)) |
2125 | dev_err(&adapter->pdev->dev, |
2126 | "payload offset should not an event number\n" ); |
2127 | return -1; |
2128 | } else { |
2129 | css = cso + skb->csum_offset; |
2130 | |
2131 | (*tpd)->word1 |= ((cso >> 1) & TPD_PLOADOFFSET_MASK) << |
2132 | TPD_PLOADOFFSET_SHIFT; |
2133 | (*tpd)->word1 |= ((css >> 1) & TPD_CCSUM_OFFSET_MASK) << |
2134 | TPD_CCSUM_OFFSET_SHIFT; |
2135 | (*tpd)->word1 |= 1 << TPD_CCSUM_EN_SHIFT; |
2136 | } |
2137 | } |
2138 | return 0; |
2139 | } |
2140 | |
2141 | static void atl1c_tx_rollback(struct atl1c_adapter *adpt, |
2142 | struct atl1c_tpd_desc *first_tpd, |
2143 | u32 queue) |
2144 | { |
2145 | struct atl1c_tpd_ring *tpd_ring = &adpt->tpd_ring[queue]; |
2146 | struct atl1c_buffer *buffer_info; |
2147 | struct atl1c_tpd_desc *tpd; |
2148 | u16 first_index, index; |
2149 | |
2150 | first_index = first_tpd - (struct atl1c_tpd_desc *)tpd_ring->desc; |
2151 | index = first_index; |
2152 | while (index != tpd_ring->next_to_use) { |
2153 | tpd = ATL1C_TPD_DESC(tpd_ring, index); |
2154 | buffer_info = &tpd_ring->buffer_info[index]; |
2155 | atl1c_clean_buffer(pdev: adpt->pdev, buffer_info, budget: 0); |
2156 | memset(tpd, 0, sizeof(struct atl1c_tpd_desc)); |
2157 | if (++index == tpd_ring->count) |
2158 | index = 0; |
2159 | } |
2160 | tpd_ring->next_to_use = first_index; |
2161 | } |
2162 | |
2163 | static int atl1c_tx_map(struct atl1c_adapter *adapter, |
2164 | struct sk_buff *skb, struct atl1c_tpd_desc *tpd, |
2165 | u32 queue) |
2166 | { |
2167 | struct atl1c_tpd_desc *use_tpd = NULL; |
2168 | struct atl1c_buffer *buffer_info = NULL; |
2169 | u16 buf_len = skb_headlen(skb); |
2170 | u16 map_len = 0; |
2171 | u16 mapped_len = 0; |
2172 | u16 hdr_len = 0; |
2173 | u16 nr_frags; |
2174 | u16 f; |
2175 | int tso; |
2176 | |
2177 | nr_frags = skb_shinfo(skb)->nr_frags; |
2178 | tso = (tpd->word1 >> TPD_LSO_EN_SHIFT) & TPD_LSO_EN_MASK; |
2179 | if (tso) { |
2180 | /* TSO */ |
2181 | hdr_len = skb_tcp_all_headers(skb); |
2182 | map_len = hdr_len; |
2183 | use_tpd = tpd; |
2184 | |
2185 | buffer_info = atl1c_get_tx_buffer(adapter, tpd: use_tpd); |
2186 | buffer_info->length = map_len; |
2187 | buffer_info->dma = dma_map_single(&adapter->pdev->dev, |
2188 | skb->data, hdr_len, |
2189 | DMA_TO_DEVICE); |
2190 | if (unlikely(dma_mapping_error(&adapter->pdev->dev, buffer_info->dma))) |
2191 | goto err_dma; |
2192 | ATL1C_SET_BUFFER_STATE(buffer_info, ATL1C_BUFFER_BUSY); |
2193 | ATL1C_SET_PCIMAP_TYPE(buffer_info, ATL1C_PCIMAP_SINGLE, |
2194 | ATL1C_PCIMAP_TODEVICE); |
2195 | mapped_len += map_len; |
2196 | use_tpd->buffer_addr = cpu_to_le64(buffer_info->dma); |
2197 | use_tpd->buffer_len = cpu_to_le16(buffer_info->length); |
2198 | } |
2199 | |
2200 | if (mapped_len < buf_len) { |
2201 | /* mapped_len == 0, means we should use the first tpd, |
2202 | which is given by caller */ |
2203 | if (mapped_len == 0) |
2204 | use_tpd = tpd; |
2205 | else { |
2206 | use_tpd = atl1c_get_tpd(adapter, queue); |
2207 | memcpy(use_tpd, tpd, sizeof(struct atl1c_tpd_desc)); |
2208 | } |
2209 | buffer_info = atl1c_get_tx_buffer(adapter, tpd: use_tpd); |
2210 | buffer_info->length = buf_len - mapped_len; |
2211 | buffer_info->dma = |
2212 | dma_map_single(&adapter->pdev->dev, |
2213 | skb->data + mapped_len, |
2214 | buffer_info->length, DMA_TO_DEVICE); |
2215 | if (unlikely(dma_mapping_error(&adapter->pdev->dev, buffer_info->dma))) |
2216 | goto err_dma; |
2217 | |
2218 | ATL1C_SET_BUFFER_STATE(buffer_info, ATL1C_BUFFER_BUSY); |
2219 | ATL1C_SET_PCIMAP_TYPE(buffer_info, ATL1C_PCIMAP_SINGLE, |
2220 | ATL1C_PCIMAP_TODEVICE); |
2221 | use_tpd->buffer_addr = cpu_to_le64(buffer_info->dma); |
2222 | use_tpd->buffer_len = cpu_to_le16(buffer_info->length); |
2223 | } |
2224 | |
2225 | for (f = 0; f < nr_frags; f++) { |
2226 | skb_frag_t *frag = &skb_shinfo(skb)->frags[f]; |
2227 | |
2228 | use_tpd = atl1c_get_tpd(adapter, queue); |
2229 | memcpy(use_tpd, tpd, sizeof(struct atl1c_tpd_desc)); |
2230 | |
2231 | buffer_info = atl1c_get_tx_buffer(adapter, tpd: use_tpd); |
2232 | buffer_info->length = skb_frag_size(frag); |
2233 | buffer_info->dma = skb_frag_dma_map(dev: &adapter->pdev->dev, |
2234 | frag, offset: 0, |
2235 | size: buffer_info->length, |
2236 | dir: DMA_TO_DEVICE); |
2237 | if (dma_mapping_error(dev: &adapter->pdev->dev, dma_addr: buffer_info->dma)) |
2238 | goto err_dma; |
2239 | |
2240 | ATL1C_SET_BUFFER_STATE(buffer_info, ATL1C_BUFFER_BUSY); |
2241 | ATL1C_SET_PCIMAP_TYPE(buffer_info, ATL1C_PCIMAP_PAGE, |
2242 | ATL1C_PCIMAP_TODEVICE); |
2243 | use_tpd->buffer_addr = cpu_to_le64(buffer_info->dma); |
2244 | use_tpd->buffer_len = cpu_to_le16(buffer_info->length); |
2245 | } |
2246 | |
2247 | /* The last tpd */ |
2248 | use_tpd->word1 |= 1 << TPD_EOP_SHIFT; |
2249 | /* The last buffer info contain the skb address, |
2250 | so it will be free after unmap */ |
2251 | buffer_info->skb = skb; |
2252 | |
2253 | return 0; |
2254 | |
2255 | err_dma: |
2256 | buffer_info->dma = 0; |
2257 | buffer_info->length = 0; |
2258 | return -1; |
2259 | } |
2260 | |
2261 | static void atl1c_tx_queue(struct atl1c_adapter *adapter, u32 queue) |
2262 | { |
2263 | struct atl1c_tpd_ring *tpd_ring = &adapter->tpd_ring[queue]; |
2264 | |
2265 | AT_WRITE_REGW(&adapter->hw, atl1c_qregs[queue].tpd_prod, |
2266 | tpd_ring->next_to_use); |
2267 | } |
2268 | |
2269 | static netdev_tx_t atl1c_xmit_frame(struct sk_buff *skb, |
2270 | struct net_device *netdev) |
2271 | { |
2272 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2273 | u32 queue = skb_get_queue_mapping(skb); |
2274 | struct netdev_queue *txq = netdev_get_tx_queue(dev: netdev, index: queue); |
2275 | struct atl1c_tpd_desc *tpd; |
2276 | u16 tpd_req; |
2277 | |
2278 | if (test_bit(__AT_DOWN, &adapter->flags)) { |
2279 | dev_kfree_skb_any(skb); |
2280 | return NETDEV_TX_OK; |
2281 | } |
2282 | |
2283 | tpd_req = atl1c_cal_tpd_req(skb); |
2284 | |
2285 | if (atl1c_tpd_avail(adapter, queue) < tpd_req) { |
2286 | /* no enough descriptor, just stop queue */ |
2287 | atl1c_tx_queue(adapter, queue); |
2288 | netif_tx_stop_queue(dev_queue: txq); |
2289 | return NETDEV_TX_BUSY; |
2290 | } |
2291 | |
2292 | tpd = atl1c_get_tpd(adapter, queue); |
2293 | |
2294 | /* do TSO and check sum */ |
2295 | if (atl1c_tso_csum(adapter, skb, tpd: &tpd, queue) != 0) { |
2296 | atl1c_tx_queue(adapter, queue); |
2297 | dev_kfree_skb_any(skb); |
2298 | return NETDEV_TX_OK; |
2299 | } |
2300 | |
2301 | if (unlikely(skb_vlan_tag_present(skb))) { |
2302 | u16 vlan = skb_vlan_tag_get(skb); |
2303 | __le16 tag; |
2304 | |
2305 | vlan = cpu_to_le16(vlan); |
2306 | AT_VLAN_TO_TAG(vlan, tag); |
2307 | tpd->word1 |= 1 << TPD_INS_VTAG_SHIFT; |
2308 | tpd->vlan_tag = tag; |
2309 | } |
2310 | |
2311 | if (skb_network_offset(skb) != ETH_HLEN) |
2312 | tpd->word1 |= 1 << TPD_ETH_TYPE_SHIFT; /* Ethernet frame */ |
2313 | |
2314 | if (atl1c_tx_map(adapter, skb, tpd, queue) < 0) { |
2315 | netif_info(adapter, tx_done, adapter->netdev, |
2316 | "tx-skb dropped due to dma error\n" ); |
2317 | /* roll back tpd/buffer */ |
2318 | atl1c_tx_rollback(adpt: adapter, first_tpd: tpd, queue); |
2319 | dev_kfree_skb_any(skb); |
2320 | } else { |
2321 | bool more = netdev_xmit_more(); |
2322 | |
2323 | if (__netdev_tx_sent_queue(dev_queue: txq, bytes: skb->len, xmit_more: more)) |
2324 | atl1c_tx_queue(adapter, queue); |
2325 | } |
2326 | |
2327 | return NETDEV_TX_OK; |
2328 | } |
2329 | |
2330 | static void atl1c_free_irq(struct atl1c_adapter *adapter) |
2331 | { |
2332 | struct net_device *netdev = adapter->netdev; |
2333 | |
2334 | free_irq(adapter->pdev->irq, netdev); |
2335 | |
2336 | if (adapter->have_msi) |
2337 | pci_disable_msi(dev: adapter->pdev); |
2338 | } |
2339 | |
2340 | static int atl1c_request_irq(struct atl1c_adapter *adapter) |
2341 | { |
2342 | struct pci_dev *pdev = adapter->pdev; |
2343 | struct net_device *netdev = adapter->netdev; |
2344 | int flags = 0; |
2345 | int err = 0; |
2346 | |
2347 | adapter->have_msi = true; |
2348 | err = pci_enable_msi(dev: adapter->pdev); |
2349 | if (err) { |
2350 | if (netif_msg_ifup(adapter)) |
2351 | dev_err(&pdev->dev, |
2352 | "Unable to allocate MSI interrupt Error: %d\n" , |
2353 | err); |
2354 | adapter->have_msi = false; |
2355 | } |
2356 | |
2357 | if (!adapter->have_msi) |
2358 | flags |= IRQF_SHARED; |
2359 | err = request_irq(irq: adapter->pdev->irq, handler: atl1c_intr, flags, |
2360 | name: netdev->name, dev: netdev); |
2361 | if (err) { |
2362 | if (netif_msg_ifup(adapter)) |
2363 | dev_err(&pdev->dev, |
2364 | "Unable to allocate interrupt Error: %d\n" , |
2365 | err); |
2366 | if (adapter->have_msi) |
2367 | pci_disable_msi(dev: adapter->pdev); |
2368 | return err; |
2369 | } |
2370 | if (netif_msg_ifup(adapter)) |
2371 | dev_dbg(&pdev->dev, "atl1c_request_irq OK\n" ); |
2372 | return err; |
2373 | } |
2374 | |
2375 | |
2376 | static void atl1c_reset_dma_ring(struct atl1c_adapter *adapter) |
2377 | { |
2378 | int i; |
2379 | /* release tx-pending skbs and reset tx/rx ring index */ |
2380 | for (i = 0; i < adapter->tx_queue_count; ++i) |
2381 | atl1c_clean_tx_ring(adapter, queue: i); |
2382 | for (i = 0; i < adapter->rx_queue_count; ++i) |
2383 | atl1c_clean_rx_ring(adapter, queue: i); |
2384 | } |
2385 | |
2386 | static int atl1c_up(struct atl1c_adapter *adapter) |
2387 | { |
2388 | struct net_device *netdev = adapter->netdev; |
2389 | int err; |
2390 | int i; |
2391 | |
2392 | netif_carrier_off(dev: netdev); |
2393 | |
2394 | err = atl1c_configure(adapter); |
2395 | if (unlikely(err)) |
2396 | goto err_up; |
2397 | |
2398 | err = atl1c_request_irq(adapter); |
2399 | if (unlikely(err)) |
2400 | goto err_up; |
2401 | |
2402 | atl1c_check_link_status(adapter); |
2403 | clear_bit(__AT_DOWN, addr: &adapter->flags); |
2404 | for (i = 0; i < adapter->tx_queue_count; ++i) |
2405 | napi_enable(n: &adapter->tpd_ring[i].napi); |
2406 | for (i = 0; i < adapter->rx_queue_count; ++i) |
2407 | napi_enable(n: &adapter->rrd_ring[i].napi); |
2408 | atl1c_irq_enable(adapter); |
2409 | netif_start_queue(dev: netdev); |
2410 | return err; |
2411 | |
2412 | err_up: |
2413 | for (i = 0; i < adapter->rx_queue_count; ++i) |
2414 | atl1c_clean_rx_ring(adapter, queue: i); |
2415 | return err; |
2416 | } |
2417 | |
2418 | static void atl1c_down(struct atl1c_adapter *adapter) |
2419 | { |
2420 | struct net_device *netdev = adapter->netdev; |
2421 | int i; |
2422 | |
2423 | atl1c_del_timer(adapter); |
2424 | adapter->work_event = 0; /* clear all event */ |
2425 | /* signal that we're down so the interrupt handler does not |
2426 | * reschedule our watchdog timer */ |
2427 | set_bit(__AT_DOWN, addr: &adapter->flags); |
2428 | netif_carrier_off(dev: netdev); |
2429 | for (i = 0; i < adapter->tx_queue_count; ++i) |
2430 | napi_disable(n: &adapter->tpd_ring[i].napi); |
2431 | for (i = 0; i < adapter->rx_queue_count; ++i) |
2432 | napi_disable(n: &adapter->rrd_ring[i].napi); |
2433 | atl1c_irq_disable(adapter); |
2434 | atl1c_free_irq(adapter); |
2435 | /* disable ASPM if device inactive */ |
2436 | atl1c_disable_l0s_l1(hw: &adapter->hw); |
2437 | /* reset MAC to disable all RX/TX */ |
2438 | atl1c_reset_mac(hw: &adapter->hw); |
2439 | msleep(msecs: 1); |
2440 | |
2441 | adapter->link_speed = SPEED_0; |
2442 | adapter->link_duplex = -1; |
2443 | atl1c_reset_dma_ring(adapter); |
2444 | } |
2445 | |
2446 | /** |
2447 | * atl1c_open - Called when a network interface is made active |
2448 | * @netdev: network interface device structure |
2449 | * |
2450 | * Returns 0 on success, negative value on failure |
2451 | * |
2452 | * The open entry point is called when a network interface is made |
2453 | * active by the system (IFF_UP). At this point all resources needed |
2454 | * for transmit and receive operations are allocated, the interrupt |
2455 | * handler is registered with the OS, the watchdog timer is started, |
2456 | * and the stack is notified that the interface is ready. |
2457 | */ |
2458 | static int atl1c_open(struct net_device *netdev) |
2459 | { |
2460 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2461 | int err; |
2462 | |
2463 | /* disallow open during test */ |
2464 | if (test_bit(__AT_TESTING, &adapter->flags)) |
2465 | return -EBUSY; |
2466 | |
2467 | /* allocate rx/tx dma buffer & descriptors */ |
2468 | err = atl1c_setup_ring_resources(adapter); |
2469 | if (unlikely(err)) |
2470 | return err; |
2471 | |
2472 | err = atl1c_up(adapter); |
2473 | if (unlikely(err)) |
2474 | goto err_up; |
2475 | |
2476 | return 0; |
2477 | |
2478 | err_up: |
2479 | atl1c_free_irq(adapter); |
2480 | atl1c_free_ring_resources(adapter); |
2481 | atl1c_reset_mac(hw: &adapter->hw); |
2482 | return err; |
2483 | } |
2484 | |
2485 | /** |
2486 | * atl1c_close - Disables a network interface |
2487 | * @netdev: network interface device structure |
2488 | * |
2489 | * Returns 0, this is not allowed to fail |
2490 | * |
2491 | * The close entry point is called when an interface is de-activated |
2492 | * by the OS. The hardware is still under the drivers control, but |
2493 | * needs to be disabled. A global MAC reset is issued to stop the |
2494 | * hardware, and all transmit and receive resources are freed. |
2495 | */ |
2496 | static int atl1c_close(struct net_device *netdev) |
2497 | { |
2498 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2499 | |
2500 | WARN_ON(test_bit(__AT_RESETTING, &adapter->flags)); |
2501 | set_bit(__AT_DOWN, addr: &adapter->flags); |
2502 | cancel_work_sync(work: &adapter->common_task); |
2503 | atl1c_down(adapter); |
2504 | atl1c_free_ring_resources(adapter); |
2505 | return 0; |
2506 | } |
2507 | |
2508 | static int atl1c_suspend(struct device *dev) |
2509 | { |
2510 | struct net_device *netdev = dev_get_drvdata(dev); |
2511 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2512 | struct atl1c_hw *hw = &adapter->hw; |
2513 | u32 wufc = adapter->wol; |
2514 | |
2515 | atl1c_disable_l0s_l1(hw); |
2516 | if (netif_running(dev: netdev)) { |
2517 | WARN_ON(test_bit(__AT_RESETTING, &adapter->flags)); |
2518 | atl1c_down(adapter); |
2519 | } |
2520 | netif_device_detach(dev: netdev); |
2521 | |
2522 | if (wufc) |
2523 | if (atl1c_phy_to_ps_link(hw) != 0) |
2524 | dev_dbg(dev, "phy power saving failed" ); |
2525 | |
2526 | atl1c_power_saving(hw, wufc); |
2527 | |
2528 | return 0; |
2529 | } |
2530 | |
2531 | #ifdef CONFIG_PM_SLEEP |
2532 | static int atl1c_resume(struct device *dev) |
2533 | { |
2534 | struct net_device *netdev = dev_get_drvdata(dev); |
2535 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2536 | |
2537 | AT_WRITE_REG(&adapter->hw, REG_WOL_CTRL, 0); |
2538 | atl1c_reset_pcie(hw: &adapter->hw, ATL1C_PCIE_L0S_L1_DISABLE); |
2539 | |
2540 | atl1c_phy_reset(hw: &adapter->hw); |
2541 | atl1c_reset_mac(hw: &adapter->hw); |
2542 | atl1c_phy_init(hw: &adapter->hw); |
2543 | |
2544 | netif_device_attach(dev: netdev); |
2545 | if (netif_running(dev: netdev)) |
2546 | atl1c_up(adapter); |
2547 | |
2548 | return 0; |
2549 | } |
2550 | #endif |
2551 | |
2552 | static void atl1c_shutdown(struct pci_dev *pdev) |
2553 | { |
2554 | struct net_device *netdev = pci_get_drvdata(pdev); |
2555 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2556 | |
2557 | atl1c_suspend(dev: &pdev->dev); |
2558 | pci_wake_from_d3(dev: pdev, enable: adapter->wol); |
2559 | pci_set_power_state(dev: pdev, PCI_D3hot); |
2560 | } |
2561 | |
2562 | static const struct net_device_ops atl1c_netdev_ops = { |
2563 | .ndo_open = atl1c_open, |
2564 | .ndo_stop = atl1c_close, |
2565 | .ndo_validate_addr = eth_validate_addr, |
2566 | .ndo_start_xmit = atl1c_xmit_frame, |
2567 | .ndo_set_mac_address = atl1c_set_mac_addr, |
2568 | .ndo_set_rx_mode = atl1c_set_multi, |
2569 | .ndo_change_mtu = atl1c_change_mtu, |
2570 | .ndo_fix_features = atl1c_fix_features, |
2571 | .ndo_set_features = atl1c_set_features, |
2572 | .ndo_eth_ioctl = atl1c_ioctl, |
2573 | .ndo_tx_timeout = atl1c_tx_timeout, |
2574 | .ndo_get_stats = atl1c_get_stats, |
2575 | #ifdef CONFIG_NET_POLL_CONTROLLER |
2576 | .ndo_poll_controller = atl1c_netpoll, |
2577 | #endif |
2578 | }; |
2579 | |
2580 | static int atl1c_init_netdev(struct net_device *netdev, struct pci_dev *pdev) |
2581 | { |
2582 | SET_NETDEV_DEV(netdev, &pdev->dev); |
2583 | pci_set_drvdata(pdev, data: netdev); |
2584 | |
2585 | netdev->netdev_ops = &atl1c_netdev_ops; |
2586 | netdev->watchdog_timeo = AT_TX_WATCHDOG; |
2587 | netdev->min_mtu = ETH_ZLEN - (ETH_HLEN + VLAN_HLEN); |
2588 | atl1c_set_ethtool_ops(netdev); |
2589 | |
2590 | /* TODO: add when ready */ |
2591 | netdev->hw_features = NETIF_F_SG | |
2592 | NETIF_F_HW_CSUM | |
2593 | NETIF_F_HW_VLAN_CTAG_RX | |
2594 | NETIF_F_TSO | |
2595 | NETIF_F_TSO6; |
2596 | netdev->features = netdev->hw_features | |
2597 | NETIF_F_HW_VLAN_CTAG_TX; |
2598 | return 0; |
2599 | } |
2600 | |
2601 | /** |
2602 | * atl1c_probe - Device Initialization Routine |
2603 | * @pdev: PCI device information struct |
2604 | * @ent: entry in atl1c_pci_tbl |
2605 | * |
2606 | * Returns 0 on success, negative on failure |
2607 | * |
2608 | * atl1c_probe initializes an adapter identified by a pci_dev structure. |
2609 | * The OS initialization, configuring of the adapter private structure, |
2610 | * and a hardware reset occur. |
2611 | */ |
2612 | static int atl1c_probe(struct pci_dev *pdev, const struct pci_device_id *ent) |
2613 | { |
2614 | struct net_device *netdev; |
2615 | struct atl1c_adapter *adapter; |
2616 | static int cards_found; |
2617 | u8 __iomem *hw_addr; |
2618 | enum atl1c_nic_type nic_type; |
2619 | u32 queue_count = 1; |
2620 | int err = 0; |
2621 | int i; |
2622 | |
2623 | /* enable device (incl. PCI PM wakeup and hotplug setup) */ |
2624 | err = pci_enable_device_mem(dev: pdev); |
2625 | if (err) |
2626 | return dev_err_probe(dev: &pdev->dev, err, fmt: "cannot enable PCI device\n" ); |
2627 | |
2628 | /* |
2629 | * The atl1c chip can DMA to 64-bit addresses, but it uses a single |
2630 | * shared register for the high 32 bits, so only a single, aligned, |
2631 | * 4 GB physical address range can be used at a time. |
2632 | * |
2633 | * Supporting 64-bit DMA on this hardware is more trouble than it's |
2634 | * worth. It is far easier to limit to 32-bit DMA than update |
2635 | * various kernel subsystems to support the mechanics required by a |
2636 | * fixed-high-32-bit system. |
2637 | */ |
2638 | err = dma_set_mask_and_coherent(dev: &pdev->dev, DMA_BIT_MASK(32)); |
2639 | if (err) { |
2640 | dev_err(&pdev->dev, "No usable DMA configuration,aborting\n" ); |
2641 | goto err_dma; |
2642 | } |
2643 | |
2644 | err = pci_request_regions(pdev, atl1c_driver_name); |
2645 | if (err) { |
2646 | dev_err(&pdev->dev, "cannot obtain PCI resources\n" ); |
2647 | goto err_pci_reg; |
2648 | } |
2649 | |
2650 | pci_set_master(dev: pdev); |
2651 | |
2652 | hw_addr = pci_ioremap_bar(pdev, bar: 0); |
2653 | if (!hw_addr) { |
2654 | err = -EIO; |
2655 | dev_err(&pdev->dev, "cannot map device registers\n" ); |
2656 | goto err_ioremap; |
2657 | } |
2658 | |
2659 | nic_type = atl1c_get_mac_type(pdev, hw_addr); |
2660 | if (nic_type == athr_mt) |
2661 | queue_count = 4; |
2662 | |
2663 | netdev = alloc_etherdev_mq(sizeof(struct atl1c_adapter), queue_count); |
2664 | if (netdev == NULL) { |
2665 | err = -ENOMEM; |
2666 | goto err_alloc_etherdev; |
2667 | } |
2668 | |
2669 | err = atl1c_init_netdev(netdev, pdev); |
2670 | if (err) { |
2671 | dev_err(&pdev->dev, "init netdevice failed\n" ); |
2672 | goto err_init_netdev; |
2673 | } |
2674 | adapter = netdev_priv(dev: netdev); |
2675 | adapter->bd_number = cards_found; |
2676 | adapter->netdev = netdev; |
2677 | adapter->pdev = pdev; |
2678 | adapter->hw.adapter = adapter; |
2679 | adapter->hw.nic_type = nic_type; |
2680 | adapter->msg_enable = netif_msg_init(debug_value: -1, default_msg_enable_bits: atl1c_default_msg); |
2681 | adapter->hw.hw_addr = hw_addr; |
2682 | adapter->tx_queue_count = queue_count; |
2683 | adapter->rx_queue_count = queue_count; |
2684 | |
2685 | /* init mii data */ |
2686 | adapter->mii.dev = netdev; |
2687 | adapter->mii.mdio_read = atl1c_mdio_read; |
2688 | adapter->mii.mdio_write = atl1c_mdio_write; |
2689 | adapter->mii.phy_id_mask = 0x1f; |
2690 | adapter->mii.reg_num_mask = MDIO_CTRL_REG_MASK; |
2691 | dev_set_threaded(dev: netdev, threaded: true); |
2692 | for (i = 0; i < adapter->rx_queue_count; ++i) |
2693 | netif_napi_add(dev: netdev, napi: &adapter->rrd_ring[i].napi, |
2694 | poll: atl1c_clean_rx); |
2695 | for (i = 0; i < adapter->tx_queue_count; ++i) |
2696 | netif_napi_add_tx(dev: netdev, napi: &adapter->tpd_ring[i].napi, |
2697 | poll: atl1c_clean_tx); |
2698 | timer_setup(&adapter->phy_config_timer, atl1c_phy_config, 0); |
2699 | /* setup the private structure */ |
2700 | err = atl1c_sw_init(adapter); |
2701 | if (err) { |
2702 | dev_err(&pdev->dev, "net device private data init failed\n" ); |
2703 | goto err_sw_init; |
2704 | } |
2705 | /* set max MTU */ |
2706 | atl1c_set_max_mtu(netdev); |
2707 | |
2708 | atl1c_reset_pcie(hw: &adapter->hw, ATL1C_PCIE_L0S_L1_DISABLE); |
2709 | |
2710 | /* Init GPHY as early as possible due to power saving issue */ |
2711 | atl1c_phy_reset(hw: &adapter->hw); |
2712 | |
2713 | err = atl1c_reset_mac(hw: &adapter->hw); |
2714 | if (err) { |
2715 | err = -EIO; |
2716 | goto err_reset; |
2717 | } |
2718 | |
2719 | /* reset the controller to |
2720 | * put the device in a known good starting state */ |
2721 | err = atl1c_phy_init(hw: &adapter->hw); |
2722 | if (err) { |
2723 | err = -EIO; |
2724 | goto err_reset; |
2725 | } |
2726 | if (atl1c_read_mac_addr(hw: &adapter->hw)) { |
2727 | /* got a random MAC address, set NET_ADDR_RANDOM to netdev */ |
2728 | netdev->addr_assign_type = NET_ADDR_RANDOM; |
2729 | } |
2730 | eth_hw_addr_set(dev: netdev, addr: adapter->hw.mac_addr); |
2731 | if (netif_msg_probe(adapter)) |
2732 | dev_dbg(&pdev->dev, "mac address : %pM\n" , |
2733 | adapter->hw.mac_addr); |
2734 | |
2735 | atl1c_hw_set_mac_addr(hw: &adapter->hw, mac_addr: adapter->hw.mac_addr); |
2736 | INIT_WORK(&adapter->common_task, atl1c_common_task); |
2737 | adapter->work_event = 0; |
2738 | err = register_netdev(dev: netdev); |
2739 | if (err) { |
2740 | dev_err(&pdev->dev, "register netdevice failed\n" ); |
2741 | goto err_register; |
2742 | } |
2743 | |
2744 | cards_found++; |
2745 | return 0; |
2746 | |
2747 | err_reset: |
2748 | err_register: |
2749 | err_sw_init: |
2750 | err_init_netdev: |
2751 | free_netdev(dev: netdev); |
2752 | err_alloc_etherdev: |
2753 | iounmap(addr: hw_addr); |
2754 | err_ioremap: |
2755 | pci_release_regions(pdev); |
2756 | err_pci_reg: |
2757 | err_dma: |
2758 | pci_disable_device(dev: pdev); |
2759 | return err; |
2760 | } |
2761 | |
2762 | /** |
2763 | * atl1c_remove - Device Removal Routine |
2764 | * @pdev: PCI device information struct |
2765 | * |
2766 | * atl1c_remove is called by the PCI subsystem to alert the driver |
2767 | * that it should release a PCI device. The could be caused by a |
2768 | * Hot-Plug event, or because the driver is going to be removed from |
2769 | * memory. |
2770 | */ |
2771 | static void atl1c_remove(struct pci_dev *pdev) |
2772 | { |
2773 | struct net_device *netdev = pci_get_drvdata(pdev); |
2774 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2775 | |
2776 | unregister_netdev(dev: netdev); |
2777 | /* restore permanent address */ |
2778 | atl1c_hw_set_mac_addr(hw: &adapter->hw, mac_addr: adapter->hw.perm_mac_addr); |
2779 | atl1c_phy_disable(hw: &adapter->hw); |
2780 | |
2781 | iounmap(addr: adapter->hw.hw_addr); |
2782 | |
2783 | pci_release_regions(pdev); |
2784 | pci_disable_device(dev: pdev); |
2785 | free_netdev(dev: netdev); |
2786 | } |
2787 | |
2788 | /** |
2789 | * atl1c_io_error_detected - called when PCI error is detected |
2790 | * @pdev: Pointer to PCI device |
2791 | * @state: The current pci connection state |
2792 | * |
2793 | * This function is called after a PCI bus error affecting |
2794 | * this device has been detected. |
2795 | */ |
2796 | static pci_ers_result_t atl1c_io_error_detected(struct pci_dev *pdev, |
2797 | pci_channel_state_t state) |
2798 | { |
2799 | struct net_device *netdev = pci_get_drvdata(pdev); |
2800 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2801 | |
2802 | netif_device_detach(dev: netdev); |
2803 | |
2804 | if (state == pci_channel_io_perm_failure) |
2805 | return PCI_ERS_RESULT_DISCONNECT; |
2806 | |
2807 | if (netif_running(dev: netdev)) |
2808 | atl1c_down(adapter); |
2809 | |
2810 | pci_disable_device(dev: pdev); |
2811 | |
2812 | /* Request a slot reset. */ |
2813 | return PCI_ERS_RESULT_NEED_RESET; |
2814 | } |
2815 | |
2816 | /** |
2817 | * atl1c_io_slot_reset - called after the pci bus has been reset. |
2818 | * @pdev: Pointer to PCI device |
2819 | * |
2820 | * Restart the card from scratch, as if from a cold-boot. Implementation |
2821 | * resembles the first-half of the e1000_resume routine. |
2822 | */ |
2823 | static pci_ers_result_t atl1c_io_slot_reset(struct pci_dev *pdev) |
2824 | { |
2825 | struct net_device *netdev = pci_get_drvdata(pdev); |
2826 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2827 | |
2828 | if (pci_enable_device(dev: pdev)) { |
2829 | if (netif_msg_hw(adapter)) |
2830 | dev_err(&pdev->dev, |
2831 | "Cannot re-enable PCI device after reset\n" ); |
2832 | return PCI_ERS_RESULT_DISCONNECT; |
2833 | } |
2834 | pci_set_master(dev: pdev); |
2835 | |
2836 | pci_enable_wake(dev: pdev, PCI_D3hot, enable: 0); |
2837 | pci_enable_wake(dev: pdev, PCI_D3cold, enable: 0); |
2838 | |
2839 | atl1c_reset_mac(hw: &adapter->hw); |
2840 | |
2841 | return PCI_ERS_RESULT_RECOVERED; |
2842 | } |
2843 | |
2844 | /** |
2845 | * atl1c_io_resume - called when traffic can start flowing again. |
2846 | * @pdev: Pointer to PCI device |
2847 | * |
2848 | * This callback is called when the error recovery driver tells us that |
2849 | * its OK to resume normal operation. Implementation resembles the |
2850 | * second-half of the atl1c_resume routine. |
2851 | */ |
2852 | static void atl1c_io_resume(struct pci_dev *pdev) |
2853 | { |
2854 | struct net_device *netdev = pci_get_drvdata(pdev); |
2855 | struct atl1c_adapter *adapter = netdev_priv(dev: netdev); |
2856 | |
2857 | if (netif_running(dev: netdev)) { |
2858 | if (atl1c_up(adapter)) { |
2859 | if (netif_msg_hw(adapter)) |
2860 | dev_err(&pdev->dev, |
2861 | "Cannot bring device back up after reset\n" ); |
2862 | return; |
2863 | } |
2864 | } |
2865 | |
2866 | netif_device_attach(dev: netdev); |
2867 | } |
2868 | |
2869 | static const struct pci_error_handlers atl1c_err_handler = { |
2870 | .error_detected = atl1c_io_error_detected, |
2871 | .slot_reset = atl1c_io_slot_reset, |
2872 | .resume = atl1c_io_resume, |
2873 | }; |
2874 | |
2875 | static SIMPLE_DEV_PM_OPS(atl1c_pm_ops, atl1c_suspend, atl1c_resume); |
2876 | |
2877 | static struct pci_driver atl1c_driver = { |
2878 | .name = atl1c_driver_name, |
2879 | .id_table = atl1c_pci_tbl, |
2880 | .probe = atl1c_probe, |
2881 | .remove = atl1c_remove, |
2882 | .shutdown = atl1c_shutdown, |
2883 | .err_handler = &atl1c_err_handler, |
2884 | .driver.pm = &atl1c_pm_ops, |
2885 | }; |
2886 | |
2887 | module_pci_driver(atl1c_driver); |
2888 | |