1 | // SPDX-License-Identifier: (GPL-2.0 OR MIT) |
2 | /* |
3 | * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates. |
4 | * stmmac XGMAC support. |
5 | */ |
6 | |
7 | #include <linux/bitrev.h> |
8 | #include <linux/crc32.h> |
9 | #include <linux/iopoll.h> |
10 | #include "stmmac.h" |
11 | #include "stmmac_ptp.h" |
12 | #include "dwxlgmac2.h" |
13 | #include "dwxgmac2.h" |
14 | |
15 | static void dwxgmac2_core_init(struct mac_device_info *hw, |
16 | struct net_device *dev) |
17 | { |
18 | void __iomem *ioaddr = hw->pcsr; |
19 | u32 tx, rx; |
20 | |
21 | tx = readl(addr: ioaddr + XGMAC_TX_CONFIG); |
22 | rx = readl(addr: ioaddr + XGMAC_RX_CONFIG); |
23 | |
24 | tx |= XGMAC_CORE_INIT_TX; |
25 | rx |= XGMAC_CORE_INIT_RX; |
26 | |
27 | if (hw->ps) { |
28 | tx |= XGMAC_CONFIG_TE; |
29 | tx &= ~hw->link.speed_mask; |
30 | |
31 | switch (hw->ps) { |
32 | case SPEED_10000: |
33 | tx |= hw->link.xgmii.speed10000; |
34 | break; |
35 | case SPEED_2500: |
36 | tx |= hw->link.speed2500; |
37 | break; |
38 | case SPEED_1000: |
39 | default: |
40 | tx |= hw->link.speed1000; |
41 | break; |
42 | } |
43 | } |
44 | |
45 | writel(val: tx, addr: ioaddr + XGMAC_TX_CONFIG); |
46 | writel(val: rx, addr: ioaddr + XGMAC_RX_CONFIG); |
47 | writel(XGMAC_INT_DEFAULT_EN, addr: ioaddr + XGMAC_INT_EN); |
48 | } |
49 | |
50 | static void dwxgmac2_set_mac(void __iomem *ioaddr, bool enable) |
51 | { |
52 | u32 tx = readl(addr: ioaddr + XGMAC_TX_CONFIG); |
53 | u32 rx = readl(addr: ioaddr + XGMAC_RX_CONFIG); |
54 | |
55 | if (enable) { |
56 | tx |= XGMAC_CONFIG_TE; |
57 | rx |= XGMAC_CONFIG_RE; |
58 | } else { |
59 | tx &= ~XGMAC_CONFIG_TE; |
60 | rx &= ~XGMAC_CONFIG_RE; |
61 | } |
62 | |
63 | writel(val: tx, addr: ioaddr + XGMAC_TX_CONFIG); |
64 | writel(val: rx, addr: ioaddr + XGMAC_RX_CONFIG); |
65 | } |
66 | |
67 | static int dwxgmac2_rx_ipc(struct mac_device_info *hw) |
68 | { |
69 | void __iomem *ioaddr = hw->pcsr; |
70 | u32 value; |
71 | |
72 | value = readl(addr: ioaddr + XGMAC_RX_CONFIG); |
73 | if (hw->rx_csum) |
74 | value |= XGMAC_CONFIG_IPC; |
75 | else |
76 | value &= ~XGMAC_CONFIG_IPC; |
77 | writel(val: value, addr: ioaddr + XGMAC_RX_CONFIG); |
78 | |
79 | return !!(readl(addr: ioaddr + XGMAC_RX_CONFIG) & XGMAC_CONFIG_IPC); |
80 | } |
81 | |
82 | static void dwxgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode, |
83 | u32 queue) |
84 | { |
85 | void __iomem *ioaddr = hw->pcsr; |
86 | u32 value; |
87 | |
88 | value = readl(addr: ioaddr + XGMAC_RXQ_CTRL0) & ~XGMAC_RXQEN(queue); |
89 | if (mode == MTL_QUEUE_AVB) |
90 | value |= 0x1 << XGMAC_RXQEN_SHIFT(queue); |
91 | else if (mode == MTL_QUEUE_DCB) |
92 | value |= 0x2 << XGMAC_RXQEN_SHIFT(queue); |
93 | writel(val: value, addr: ioaddr + XGMAC_RXQ_CTRL0); |
94 | } |
95 | |
96 | static void dwxgmac2_rx_queue_prio(struct mac_device_info *hw, u32 prio, |
97 | u32 queue) |
98 | { |
99 | void __iomem *ioaddr = hw->pcsr; |
100 | u32 clear_mask = 0; |
101 | u32 ctrl2, ctrl3; |
102 | int i; |
103 | |
104 | ctrl2 = readl(addr: ioaddr + XGMAC_RXQ_CTRL2); |
105 | ctrl3 = readl(addr: ioaddr + XGMAC_RXQ_CTRL3); |
106 | |
107 | /* The software must ensure that the same priority |
108 | * is not mapped to multiple Rx queues |
109 | */ |
110 | for (i = 0; i < 4; i++) |
111 | clear_mask |= ((prio << XGMAC_PSRQ_SHIFT(i)) & |
112 | XGMAC_PSRQ(i)); |
113 | |
114 | ctrl2 &= ~clear_mask; |
115 | ctrl3 &= ~clear_mask; |
116 | |
117 | /* First assign new priorities to a queue, then |
118 | * clear them from others queues |
119 | */ |
120 | if (queue < 4) { |
121 | ctrl2 |= (prio << XGMAC_PSRQ_SHIFT(queue)) & |
122 | XGMAC_PSRQ(queue); |
123 | |
124 | writel(val: ctrl2, addr: ioaddr + XGMAC_RXQ_CTRL2); |
125 | writel(val: ctrl3, addr: ioaddr + XGMAC_RXQ_CTRL3); |
126 | } else { |
127 | queue -= 4; |
128 | |
129 | ctrl3 |= (prio << XGMAC_PSRQ_SHIFT(queue)) & |
130 | XGMAC_PSRQ(queue); |
131 | |
132 | writel(val: ctrl3, addr: ioaddr + XGMAC_RXQ_CTRL3); |
133 | writel(val: ctrl2, addr: ioaddr + XGMAC_RXQ_CTRL2); |
134 | } |
135 | } |
136 | |
137 | static void dwxgmac2_tx_queue_prio(struct mac_device_info *hw, u32 prio, |
138 | u32 queue) |
139 | { |
140 | void __iomem *ioaddr = hw->pcsr; |
141 | u32 value, reg; |
142 | |
143 | reg = (queue < 4) ? XGMAC_TC_PRTY_MAP0 : XGMAC_TC_PRTY_MAP1; |
144 | if (queue >= 4) |
145 | queue -= 4; |
146 | |
147 | value = readl(addr: ioaddr + reg); |
148 | value &= ~XGMAC_PSTC(queue); |
149 | value |= (prio << XGMAC_PSTC_SHIFT(queue)) & XGMAC_PSTC(queue); |
150 | |
151 | writel(val: value, addr: ioaddr + reg); |
152 | } |
153 | |
154 | static void dwxgmac2_rx_queue_routing(struct mac_device_info *hw, |
155 | u8 packet, u32 queue) |
156 | { |
157 | void __iomem *ioaddr = hw->pcsr; |
158 | u32 value; |
159 | |
160 | static const struct stmmac_rx_routing dwxgmac2_route_possibilities[] = { |
161 | { XGMAC_AVCPQ, XGMAC_AVCPQ_SHIFT }, |
162 | { XGMAC_PTPQ, XGMAC_PTPQ_SHIFT }, |
163 | { XGMAC_DCBCPQ, XGMAC_DCBCPQ_SHIFT }, |
164 | { XGMAC_UPQ, XGMAC_UPQ_SHIFT }, |
165 | { XGMAC_MCBCQ, XGMAC_MCBCQ_SHIFT }, |
166 | }; |
167 | |
168 | value = readl(addr: ioaddr + XGMAC_RXQ_CTRL1); |
169 | |
170 | /* routing configuration */ |
171 | value &= ~dwxgmac2_route_possibilities[packet - 1].reg_mask; |
172 | value |= (queue << dwxgmac2_route_possibilities[packet - 1].reg_shift) & |
173 | dwxgmac2_route_possibilities[packet - 1].reg_mask; |
174 | |
175 | /* some packets require extra ops */ |
176 | if (packet == PACKET_AVCPQ) |
177 | value |= FIELD_PREP(XGMAC_TACPQE, 1); |
178 | else if (packet == PACKET_MCBCQ) |
179 | value |= FIELD_PREP(XGMAC_MCBCQEN, 1); |
180 | |
181 | writel(val: value, addr: ioaddr + XGMAC_RXQ_CTRL1); |
182 | } |
183 | |
184 | static void dwxgmac2_prog_mtl_rx_algorithms(struct mac_device_info *hw, |
185 | u32 rx_alg) |
186 | { |
187 | void __iomem *ioaddr = hw->pcsr; |
188 | u32 value; |
189 | |
190 | value = readl(addr: ioaddr + XGMAC_MTL_OPMODE); |
191 | value &= ~XGMAC_RAA; |
192 | |
193 | switch (rx_alg) { |
194 | case MTL_RX_ALGORITHM_SP: |
195 | break; |
196 | case MTL_RX_ALGORITHM_WSP: |
197 | value |= XGMAC_RAA; |
198 | break; |
199 | default: |
200 | break; |
201 | } |
202 | |
203 | writel(val: value, addr: ioaddr + XGMAC_MTL_OPMODE); |
204 | } |
205 | |
206 | static void dwxgmac2_prog_mtl_tx_algorithms(struct mac_device_info *hw, |
207 | u32 tx_alg) |
208 | { |
209 | void __iomem *ioaddr = hw->pcsr; |
210 | bool ets = true; |
211 | u32 value; |
212 | int i; |
213 | |
214 | value = readl(addr: ioaddr + XGMAC_MTL_OPMODE); |
215 | value &= ~XGMAC_ETSALG; |
216 | |
217 | switch (tx_alg) { |
218 | case MTL_TX_ALGORITHM_WRR: |
219 | value |= XGMAC_WRR; |
220 | break; |
221 | case MTL_TX_ALGORITHM_WFQ: |
222 | value |= XGMAC_WFQ; |
223 | break; |
224 | case MTL_TX_ALGORITHM_DWRR: |
225 | value |= XGMAC_DWRR; |
226 | break; |
227 | default: |
228 | ets = false; |
229 | break; |
230 | } |
231 | |
232 | writel(val: value, addr: ioaddr + XGMAC_MTL_OPMODE); |
233 | |
234 | /* Set ETS if desired */ |
235 | for (i = 0; i < MTL_MAX_TX_QUEUES; i++) { |
236 | value = readl(addr: ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(i)); |
237 | value &= ~XGMAC_TSA; |
238 | if (ets) |
239 | value |= XGMAC_ETS; |
240 | writel(val: value, addr: ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(i)); |
241 | } |
242 | } |
243 | |
244 | static void dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv *priv, |
245 | struct mac_device_info *hw, |
246 | u32 weight, u32 queue) |
247 | { |
248 | void __iomem *ioaddr = hw->pcsr; |
249 | |
250 | writel(val: weight, addr: ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue)); |
251 | } |
252 | |
253 | static void dwxgmac2_map_mtl_to_dma(struct mac_device_info *hw, u32 queue, |
254 | u32 chan) |
255 | { |
256 | void __iomem *ioaddr = hw->pcsr; |
257 | u32 value, reg; |
258 | |
259 | reg = (queue < 4) ? XGMAC_MTL_RXQ_DMA_MAP0 : XGMAC_MTL_RXQ_DMA_MAP1; |
260 | if (queue >= 4) |
261 | queue -= 4; |
262 | |
263 | value = readl(addr: ioaddr + reg); |
264 | value &= ~XGMAC_QxMDMACH(queue); |
265 | value |= (chan << XGMAC_QxMDMACH_SHIFT(queue)) & XGMAC_QxMDMACH(queue); |
266 | |
267 | writel(val: value, addr: ioaddr + reg); |
268 | } |
269 | |
270 | static void dwxgmac2_config_cbs(struct stmmac_priv *priv, |
271 | struct mac_device_info *hw, |
272 | u32 send_slope, u32 idle_slope, |
273 | u32 high_credit, u32 low_credit, u32 queue) |
274 | { |
275 | void __iomem *ioaddr = hw->pcsr; |
276 | u32 value; |
277 | |
278 | writel(val: send_slope, addr: ioaddr + XGMAC_MTL_TCx_SENDSLOPE(queue)); |
279 | writel(val: idle_slope, addr: ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue)); |
280 | writel(val: high_credit, addr: ioaddr + XGMAC_MTL_TCx_HICREDIT(queue)); |
281 | writel(val: low_credit, addr: ioaddr + XGMAC_MTL_TCx_LOCREDIT(queue)); |
282 | |
283 | value = readl(addr: ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue)); |
284 | value &= ~XGMAC_TSA; |
285 | value |= XGMAC_CC | XGMAC_CBS; |
286 | writel(val: value, addr: ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue)); |
287 | } |
288 | |
289 | static void dwxgmac2_dump_regs(struct mac_device_info *hw, u32 *reg_space) |
290 | { |
291 | void __iomem *ioaddr = hw->pcsr; |
292 | int i; |
293 | |
294 | for (i = 0; i < XGMAC_MAC_REGSIZE; i++) |
295 | reg_space[i] = readl(addr: ioaddr + i * 4); |
296 | } |
297 | |
298 | static int dwxgmac2_host_irq_status(struct mac_device_info *hw, |
299 | struct stmmac_extra_stats *x) |
300 | { |
301 | void __iomem *ioaddr = hw->pcsr; |
302 | u32 stat, en; |
303 | int ret = 0; |
304 | |
305 | en = readl(addr: ioaddr + XGMAC_INT_EN); |
306 | stat = readl(addr: ioaddr + XGMAC_INT_STATUS); |
307 | |
308 | stat &= en; |
309 | |
310 | if (stat & XGMAC_PMTIS) { |
311 | x->irq_receive_pmt_irq_n++; |
312 | readl(addr: ioaddr + XGMAC_PMT); |
313 | } |
314 | |
315 | if (stat & XGMAC_LPIIS) { |
316 | u32 lpi = readl(addr: ioaddr + XGMAC_LPI_CTRL); |
317 | |
318 | if (lpi & XGMAC_TLPIEN) { |
319 | ret |= CORE_IRQ_TX_PATH_IN_LPI_MODE; |
320 | x->irq_tx_path_in_lpi_mode_n++; |
321 | } |
322 | if (lpi & XGMAC_TLPIEX) { |
323 | ret |= CORE_IRQ_TX_PATH_EXIT_LPI_MODE; |
324 | x->irq_tx_path_exit_lpi_mode_n++; |
325 | } |
326 | if (lpi & XGMAC_RLPIEN) |
327 | x->irq_rx_path_in_lpi_mode_n++; |
328 | if (lpi & XGMAC_RLPIEX) |
329 | x->irq_rx_path_exit_lpi_mode_n++; |
330 | } |
331 | |
332 | return ret; |
333 | } |
334 | |
335 | static int dwxgmac2_host_mtl_irq_status(struct stmmac_priv *priv, |
336 | struct mac_device_info *hw, u32 chan) |
337 | { |
338 | void __iomem *ioaddr = hw->pcsr; |
339 | int ret = 0; |
340 | u32 status; |
341 | |
342 | status = readl(addr: ioaddr + XGMAC_MTL_INT_STATUS); |
343 | if (status & BIT(chan)) { |
344 | u32 chan_status = readl(addr: ioaddr + XGMAC_MTL_QINT_STATUS(chan)); |
345 | |
346 | if (chan_status & XGMAC_RXOVFIS) |
347 | ret |= CORE_IRQ_MTL_RX_OVERFLOW; |
348 | |
349 | writel(val: ~0x0, addr: ioaddr + XGMAC_MTL_QINT_STATUS(chan)); |
350 | } |
351 | |
352 | return ret; |
353 | } |
354 | |
355 | static void dwxgmac2_flow_ctrl(struct mac_device_info *hw, unsigned int duplex, |
356 | unsigned int fc, unsigned int pause_time, |
357 | u32 tx_cnt) |
358 | { |
359 | void __iomem *ioaddr = hw->pcsr; |
360 | u32 i; |
361 | |
362 | if (fc & FLOW_RX) |
363 | writel(XGMAC_RFE, addr: ioaddr + XGMAC_RX_FLOW_CTRL); |
364 | if (fc & FLOW_TX) { |
365 | for (i = 0; i < tx_cnt; i++) { |
366 | u32 value = XGMAC_TFE; |
367 | |
368 | if (duplex) |
369 | value |= pause_time << XGMAC_PT_SHIFT; |
370 | |
371 | writel(val: value, addr: ioaddr + XGMAC_Qx_TX_FLOW_CTRL(i)); |
372 | } |
373 | } |
374 | } |
375 | |
376 | static void dwxgmac2_pmt(struct mac_device_info *hw, unsigned long mode) |
377 | { |
378 | void __iomem *ioaddr = hw->pcsr; |
379 | u32 val = 0x0; |
380 | |
381 | if (mode & WAKE_MAGIC) |
382 | val |= XGMAC_PWRDWN | XGMAC_MGKPKTEN; |
383 | if (mode & WAKE_UCAST) |
384 | val |= XGMAC_PWRDWN | XGMAC_GLBLUCAST | XGMAC_RWKPKTEN; |
385 | if (val) { |
386 | u32 cfg = readl(addr: ioaddr + XGMAC_RX_CONFIG); |
387 | cfg |= XGMAC_CONFIG_RE; |
388 | writel(val: cfg, addr: ioaddr + XGMAC_RX_CONFIG); |
389 | } |
390 | |
391 | writel(val, addr: ioaddr + XGMAC_PMT); |
392 | } |
393 | |
394 | static void dwxgmac2_set_umac_addr(struct mac_device_info *hw, |
395 | const unsigned char *addr, |
396 | unsigned int reg_n) |
397 | { |
398 | void __iomem *ioaddr = hw->pcsr; |
399 | u32 value; |
400 | |
401 | value = (addr[5] << 8) | addr[4]; |
402 | writel(val: value | XGMAC_AE, addr: ioaddr + XGMAC_ADDRx_HIGH(reg_n)); |
403 | |
404 | value = (addr[3] << 24) | (addr[2] << 16) | (addr[1] << 8) | addr[0]; |
405 | writel(val: value, addr: ioaddr + XGMAC_ADDRx_LOW(reg_n)); |
406 | } |
407 | |
408 | static void dwxgmac2_get_umac_addr(struct mac_device_info *hw, |
409 | unsigned char *addr, unsigned int reg_n) |
410 | { |
411 | void __iomem *ioaddr = hw->pcsr; |
412 | u32 hi_addr, lo_addr; |
413 | |
414 | /* Read the MAC address from the hardware */ |
415 | hi_addr = readl(addr: ioaddr + XGMAC_ADDRx_HIGH(reg_n)); |
416 | lo_addr = readl(addr: ioaddr + XGMAC_ADDRx_LOW(reg_n)); |
417 | |
418 | /* Extract the MAC address from the high and low words */ |
419 | addr[0] = lo_addr & 0xff; |
420 | addr[1] = (lo_addr >> 8) & 0xff; |
421 | addr[2] = (lo_addr >> 16) & 0xff; |
422 | addr[3] = (lo_addr >> 24) & 0xff; |
423 | addr[4] = hi_addr & 0xff; |
424 | addr[5] = (hi_addr >> 8) & 0xff; |
425 | } |
426 | |
427 | static void dwxgmac2_set_eee_mode(struct mac_device_info *hw, |
428 | bool en_tx_lpi_clockgating) |
429 | { |
430 | void __iomem *ioaddr = hw->pcsr; |
431 | u32 value; |
432 | |
433 | value = readl(addr: ioaddr + XGMAC_LPI_CTRL); |
434 | |
435 | value |= XGMAC_LPITXEN | XGMAC_LPITXA; |
436 | if (en_tx_lpi_clockgating) |
437 | value |= XGMAC_TXCGE; |
438 | |
439 | writel(val: value, addr: ioaddr + XGMAC_LPI_CTRL); |
440 | } |
441 | |
442 | static void dwxgmac2_reset_eee_mode(struct mac_device_info *hw) |
443 | { |
444 | void __iomem *ioaddr = hw->pcsr; |
445 | u32 value; |
446 | |
447 | value = readl(addr: ioaddr + XGMAC_LPI_CTRL); |
448 | value &= ~(XGMAC_LPITXEN | XGMAC_LPITXA | XGMAC_TXCGE); |
449 | writel(val: value, addr: ioaddr + XGMAC_LPI_CTRL); |
450 | } |
451 | |
452 | static void dwxgmac2_set_eee_pls(struct mac_device_info *hw, int link) |
453 | { |
454 | void __iomem *ioaddr = hw->pcsr; |
455 | u32 value; |
456 | |
457 | value = readl(addr: ioaddr + XGMAC_LPI_CTRL); |
458 | if (link) |
459 | value |= XGMAC_PLS; |
460 | else |
461 | value &= ~XGMAC_PLS; |
462 | writel(val: value, addr: ioaddr + XGMAC_LPI_CTRL); |
463 | } |
464 | |
465 | static void dwxgmac2_set_eee_timer(struct mac_device_info *hw, int ls, int tw) |
466 | { |
467 | void __iomem *ioaddr = hw->pcsr; |
468 | u32 value; |
469 | |
470 | value = (tw & 0xffff) | ((ls & 0x3ff) << 16); |
471 | writel(val: value, addr: ioaddr + XGMAC_LPI_TIMER_CTRL); |
472 | } |
473 | |
474 | static void dwxgmac2_set_mchash(void __iomem *ioaddr, u32 *mcfilterbits, |
475 | int mcbitslog2) |
476 | { |
477 | int numhashregs, regs; |
478 | |
479 | switch (mcbitslog2) { |
480 | case 6: |
481 | numhashregs = 2; |
482 | break; |
483 | case 7: |
484 | numhashregs = 4; |
485 | break; |
486 | case 8: |
487 | numhashregs = 8; |
488 | break; |
489 | default: |
490 | return; |
491 | } |
492 | |
493 | for (regs = 0; regs < numhashregs; regs++) |
494 | writel(val: mcfilterbits[regs], addr: ioaddr + XGMAC_HASH_TABLE(regs)); |
495 | } |
496 | |
497 | static void dwxgmac2_set_filter(struct mac_device_info *hw, |
498 | struct net_device *dev) |
499 | { |
500 | void __iomem *ioaddr = (void __iomem *)dev->base_addr; |
501 | u32 value = readl(addr: ioaddr + XGMAC_PACKET_FILTER); |
502 | int mcbitslog2 = hw->mcast_bits_log2; |
503 | u32 mc_filter[8]; |
504 | int i; |
505 | |
506 | value &= ~(XGMAC_FILTER_PR | XGMAC_FILTER_HMC | XGMAC_FILTER_PM); |
507 | value |= XGMAC_FILTER_HPF; |
508 | |
509 | memset(mc_filter, 0, sizeof(mc_filter)); |
510 | |
511 | if (dev->flags & IFF_PROMISC) { |
512 | value |= XGMAC_FILTER_PR; |
513 | value |= XGMAC_FILTER_PCF; |
514 | } else if ((dev->flags & IFF_ALLMULTI) || |
515 | (netdev_mc_count(dev) > hw->multicast_filter_bins)) { |
516 | value |= XGMAC_FILTER_PM; |
517 | |
518 | for (i = 0; i < XGMAC_MAX_HASH_TABLE; i++) |
519 | writel(val: ~0x0, addr: ioaddr + XGMAC_HASH_TABLE(i)); |
520 | } else if (!netdev_mc_empty(dev) && (dev->flags & IFF_MULTICAST)) { |
521 | struct netdev_hw_addr *ha; |
522 | |
523 | value |= XGMAC_FILTER_HMC; |
524 | |
525 | netdev_for_each_mc_addr(ha, dev) { |
526 | u32 nr = (bitrev32(~crc32_le(~0, ha->addr, 6)) >> |
527 | (32 - mcbitslog2)); |
528 | mc_filter[nr >> 5] |= (1 << (nr & 0x1F)); |
529 | } |
530 | } |
531 | |
532 | dwxgmac2_set_mchash(ioaddr, mcfilterbits: mc_filter, mcbitslog2); |
533 | |
534 | /* Handle multiple unicast addresses */ |
535 | if (netdev_uc_count(dev) > hw->unicast_filter_entries) { |
536 | value |= XGMAC_FILTER_PR; |
537 | } else { |
538 | struct netdev_hw_addr *ha; |
539 | int reg = 1; |
540 | |
541 | netdev_for_each_uc_addr(ha, dev) { |
542 | dwxgmac2_set_umac_addr(hw, addr: ha->addr, reg_n: reg); |
543 | reg++; |
544 | } |
545 | |
546 | for ( ; reg < XGMAC_ADDR_MAX; reg++) { |
547 | writel(val: 0, addr: ioaddr + XGMAC_ADDRx_HIGH(reg)); |
548 | writel(val: 0, addr: ioaddr + XGMAC_ADDRx_LOW(reg)); |
549 | } |
550 | } |
551 | |
552 | writel(val: value, addr: ioaddr + XGMAC_PACKET_FILTER); |
553 | } |
554 | |
555 | static void dwxgmac2_set_mac_loopback(void __iomem *ioaddr, bool enable) |
556 | { |
557 | u32 value = readl(addr: ioaddr + XGMAC_RX_CONFIG); |
558 | |
559 | if (enable) |
560 | value |= XGMAC_CONFIG_LM; |
561 | else |
562 | value &= ~XGMAC_CONFIG_LM; |
563 | |
564 | writel(val: value, addr: ioaddr + XGMAC_RX_CONFIG); |
565 | } |
566 | |
567 | static int (void __iomem *ioaddr, bool is_key, int idx, |
568 | u32 val) |
569 | { |
570 | u32 ctrl = 0; |
571 | |
572 | writel(val, addr: ioaddr + XGMAC_RSS_DATA); |
573 | ctrl |= idx << XGMAC_RSSIA_SHIFT; |
574 | ctrl |= is_key ? XGMAC_ADDRT : 0x0; |
575 | ctrl |= XGMAC_OB; |
576 | writel(val: ctrl, addr: ioaddr + XGMAC_RSS_ADDR); |
577 | |
578 | return readl_poll_timeout(ioaddr + XGMAC_RSS_ADDR, ctrl, |
579 | !(ctrl & XGMAC_OB), 100, 10000); |
580 | } |
581 | |
582 | static int (struct mac_device_info *hw, |
583 | struct stmmac_rss *cfg, u32 num_rxq) |
584 | { |
585 | void __iomem *ioaddr = hw->pcsr; |
586 | u32 value, *key; |
587 | int i, ret; |
588 | |
589 | value = readl(addr: ioaddr + XGMAC_RSS_CTRL); |
590 | if (!cfg || !cfg->enable) { |
591 | value &= ~XGMAC_RSSE; |
592 | writel(val: value, addr: ioaddr + XGMAC_RSS_CTRL); |
593 | return 0; |
594 | } |
595 | |
596 | key = (u32 *)cfg->key; |
597 | for (i = 0; i < (ARRAY_SIZE(cfg->key) / sizeof(u32)); i++) { |
598 | ret = dwxgmac2_rss_write_reg(ioaddr, is_key: true, idx: i, val: key[i]); |
599 | if (ret) |
600 | return ret; |
601 | } |
602 | |
603 | for (i = 0; i < ARRAY_SIZE(cfg->table); i++) { |
604 | ret = dwxgmac2_rss_write_reg(ioaddr, is_key: false, idx: i, val: cfg->table[i]); |
605 | if (ret) |
606 | return ret; |
607 | } |
608 | |
609 | for (i = 0; i < num_rxq; i++) |
610 | dwxgmac2_map_mtl_to_dma(hw, queue: i, XGMAC_QDDMACH); |
611 | |
612 | value |= XGMAC_UDP4TE | XGMAC_TCP4TE | XGMAC_IP2TE | XGMAC_RSSE; |
613 | writel(val: value, addr: ioaddr + XGMAC_RSS_CTRL); |
614 | return 0; |
615 | } |
616 | |
617 | static void dwxgmac2_update_vlan_hash(struct mac_device_info *hw, u32 hash, |
618 | __le16 perfect_match, bool is_double) |
619 | { |
620 | void __iomem *ioaddr = hw->pcsr; |
621 | |
622 | writel(val: hash, addr: ioaddr + XGMAC_VLAN_HASH_TABLE); |
623 | |
624 | if (hash) { |
625 | u32 value = readl(addr: ioaddr + XGMAC_PACKET_FILTER); |
626 | |
627 | value |= XGMAC_FILTER_VTFE; |
628 | |
629 | writel(val: value, addr: ioaddr + XGMAC_PACKET_FILTER); |
630 | |
631 | value = readl(addr: ioaddr + XGMAC_VLAN_TAG); |
632 | |
633 | value |= XGMAC_VLAN_VTHM | XGMAC_VLAN_ETV; |
634 | if (is_double) { |
635 | value |= XGMAC_VLAN_EDVLP; |
636 | value |= XGMAC_VLAN_ESVL; |
637 | value |= XGMAC_VLAN_DOVLTC; |
638 | } else { |
639 | value &= ~XGMAC_VLAN_EDVLP; |
640 | value &= ~XGMAC_VLAN_ESVL; |
641 | value &= ~XGMAC_VLAN_DOVLTC; |
642 | } |
643 | |
644 | value &= ~XGMAC_VLAN_VID; |
645 | writel(val: value, addr: ioaddr + XGMAC_VLAN_TAG); |
646 | } else if (perfect_match) { |
647 | u32 value = readl(addr: ioaddr + XGMAC_PACKET_FILTER); |
648 | |
649 | value |= XGMAC_FILTER_VTFE; |
650 | |
651 | writel(val: value, addr: ioaddr + XGMAC_PACKET_FILTER); |
652 | |
653 | value = readl(addr: ioaddr + XGMAC_VLAN_TAG); |
654 | |
655 | value &= ~XGMAC_VLAN_VTHM; |
656 | value |= XGMAC_VLAN_ETV; |
657 | if (is_double) { |
658 | value |= XGMAC_VLAN_EDVLP; |
659 | value |= XGMAC_VLAN_ESVL; |
660 | value |= XGMAC_VLAN_DOVLTC; |
661 | } else { |
662 | value &= ~XGMAC_VLAN_EDVLP; |
663 | value &= ~XGMAC_VLAN_ESVL; |
664 | value &= ~XGMAC_VLAN_DOVLTC; |
665 | } |
666 | |
667 | value &= ~XGMAC_VLAN_VID; |
668 | writel(val: value | perfect_match, addr: ioaddr + XGMAC_VLAN_TAG); |
669 | } else { |
670 | u32 value = readl(addr: ioaddr + XGMAC_PACKET_FILTER); |
671 | |
672 | value &= ~XGMAC_FILTER_VTFE; |
673 | |
674 | writel(val: value, addr: ioaddr + XGMAC_PACKET_FILTER); |
675 | |
676 | value = readl(addr: ioaddr + XGMAC_VLAN_TAG); |
677 | |
678 | value &= ~(XGMAC_VLAN_VTHM | XGMAC_VLAN_ETV); |
679 | value &= ~(XGMAC_VLAN_EDVLP | XGMAC_VLAN_ESVL); |
680 | value &= ~XGMAC_VLAN_DOVLTC; |
681 | value &= ~XGMAC_VLAN_VID; |
682 | |
683 | writel(val: value, addr: ioaddr + XGMAC_VLAN_TAG); |
684 | } |
685 | } |
686 | |
687 | struct dwxgmac3_error_desc { |
688 | bool valid; |
689 | const char *desc; |
690 | const char *detailed_desc; |
691 | }; |
692 | |
693 | #define STAT_OFF(field) offsetof(struct stmmac_safety_stats, field) |
694 | |
695 | static void dwxgmac3_log_error(struct net_device *ndev, u32 value, bool corr, |
696 | const char *module_name, |
697 | const struct dwxgmac3_error_desc *desc, |
698 | unsigned long field_offset, |
699 | struct stmmac_safety_stats *stats) |
700 | { |
701 | unsigned long loc, mask; |
702 | u8 *bptr = (u8 *)stats; |
703 | unsigned long *ptr; |
704 | |
705 | ptr = (unsigned long *)(bptr + field_offset); |
706 | |
707 | mask = value; |
708 | for_each_set_bit(loc, &mask, 32) { |
709 | netdev_err(dev: ndev, format: "Found %s error in %s: '%s: %s'\n" , corr ? |
710 | "correctable" : "uncorrectable" , module_name, |
711 | desc[loc].desc, desc[loc].detailed_desc); |
712 | |
713 | /* Update counters */ |
714 | ptr[loc]++; |
715 | } |
716 | } |
717 | |
718 | static const struct dwxgmac3_error_desc dwxgmac3_mac_errors[32]= { |
719 | { true, "ATPES" , "Application Transmit Interface Parity Check Error" }, |
720 | { true, "DPES" , "Descriptor Cache Data Path Parity Check Error" }, |
721 | { true, "TPES" , "TSO Data Path Parity Check Error" }, |
722 | { true, "TSOPES" , "TSO Header Data Path Parity Check Error" }, |
723 | { true, "MTPES" , "MTL Data Path Parity Check Error" }, |
724 | { true, "MTSPES" , "MTL TX Status Data Path Parity Check Error" }, |
725 | { true, "MTBUPES" , "MAC TBU Data Path Parity Check Error" }, |
726 | { true, "MTFCPES" , "MAC TFC Data Path Parity Check Error" }, |
727 | { true, "ARPES" , "Application Receive Interface Data Path Parity Check Error" }, |
728 | { true, "MRWCPES" , "MTL RWC Data Path Parity Check Error" }, |
729 | { true, "MRRCPES" , "MTL RCC Data Path Parity Check Error" }, |
730 | { true, "CWPES" , "CSR Write Data Path Parity Check Error" }, |
731 | { true, "ASRPES" , "AXI Slave Read Data Path Parity Check Error" }, |
732 | { true, "TTES" , "TX FSM Timeout Error" }, |
733 | { true, "RTES" , "RX FSM Timeout Error" }, |
734 | { true, "CTES" , "CSR FSM Timeout Error" }, |
735 | { true, "ATES" , "APP FSM Timeout Error" }, |
736 | { true, "PTES" , "PTP FSM Timeout Error" }, |
737 | { false, "UNKNOWN" , "Unknown Error" }, /* 18 */ |
738 | { false, "UNKNOWN" , "Unknown Error" }, /* 19 */ |
739 | { false, "UNKNOWN" , "Unknown Error" }, /* 20 */ |
740 | { true, "MSTTES" , "Master Read/Write Timeout Error" }, |
741 | { true, "SLVTES" , "Slave Read/Write Timeout Error" }, |
742 | { true, "ATITES" , "Application Timeout on ATI Interface Error" }, |
743 | { true, "ARITES" , "Application Timeout on ARI Interface Error" }, |
744 | { true, "FSMPES" , "FSM State Parity Error" }, |
745 | { false, "UNKNOWN" , "Unknown Error" }, /* 26 */ |
746 | { false, "UNKNOWN" , "Unknown Error" }, /* 27 */ |
747 | { false, "UNKNOWN" , "Unknown Error" }, /* 28 */ |
748 | { false, "UNKNOWN" , "Unknown Error" }, /* 29 */ |
749 | { false, "UNKNOWN" , "Unknown Error" }, /* 30 */ |
750 | { true, "CPI" , "Control Register Parity Check Error" }, |
751 | }; |
752 | |
753 | static void dwxgmac3_handle_mac_err(struct net_device *ndev, |
754 | void __iomem *ioaddr, bool correctable, |
755 | struct stmmac_safety_stats *stats) |
756 | { |
757 | u32 value; |
758 | |
759 | value = readl(addr: ioaddr + XGMAC_MAC_DPP_FSM_INT_STATUS); |
760 | writel(val: value, addr: ioaddr + XGMAC_MAC_DPP_FSM_INT_STATUS); |
761 | |
762 | dwxgmac3_log_error(ndev, value, corr: correctable, module_name: "MAC" , |
763 | desc: dwxgmac3_mac_errors, STAT_OFF(mac_errors), stats); |
764 | } |
765 | |
766 | static const struct dwxgmac3_error_desc dwxgmac3_mtl_errors[32]= { |
767 | { true, "TXCES" , "MTL TX Memory Error" }, |
768 | { true, "TXAMS" , "MTL TX Memory Address Mismatch Error" }, |
769 | { true, "TXUES" , "MTL TX Memory Error" }, |
770 | { false, "UNKNOWN" , "Unknown Error" }, /* 3 */ |
771 | { true, "RXCES" , "MTL RX Memory Error" }, |
772 | { true, "RXAMS" , "MTL RX Memory Address Mismatch Error" }, |
773 | { true, "RXUES" , "MTL RX Memory Error" }, |
774 | { false, "UNKNOWN" , "Unknown Error" }, /* 7 */ |
775 | { true, "ECES" , "MTL EST Memory Error" }, |
776 | { true, "EAMS" , "MTL EST Memory Address Mismatch Error" }, |
777 | { true, "EUES" , "MTL EST Memory Error" }, |
778 | { false, "UNKNOWN" , "Unknown Error" }, /* 11 */ |
779 | { true, "RPCES" , "MTL RX Parser Memory Error" }, |
780 | { true, "RPAMS" , "MTL RX Parser Memory Address Mismatch Error" }, |
781 | { true, "RPUES" , "MTL RX Parser Memory Error" }, |
782 | { false, "UNKNOWN" , "Unknown Error" }, /* 15 */ |
783 | { false, "UNKNOWN" , "Unknown Error" }, /* 16 */ |
784 | { false, "UNKNOWN" , "Unknown Error" }, /* 17 */ |
785 | { false, "UNKNOWN" , "Unknown Error" }, /* 18 */ |
786 | { false, "UNKNOWN" , "Unknown Error" }, /* 19 */ |
787 | { false, "UNKNOWN" , "Unknown Error" }, /* 20 */ |
788 | { false, "UNKNOWN" , "Unknown Error" }, /* 21 */ |
789 | { false, "UNKNOWN" , "Unknown Error" }, /* 22 */ |
790 | { false, "UNKNOWN" , "Unknown Error" }, /* 23 */ |
791 | { false, "UNKNOWN" , "Unknown Error" }, /* 24 */ |
792 | { false, "UNKNOWN" , "Unknown Error" }, /* 25 */ |
793 | { false, "UNKNOWN" , "Unknown Error" }, /* 26 */ |
794 | { false, "UNKNOWN" , "Unknown Error" }, /* 27 */ |
795 | { false, "UNKNOWN" , "Unknown Error" }, /* 28 */ |
796 | { false, "UNKNOWN" , "Unknown Error" }, /* 29 */ |
797 | { false, "UNKNOWN" , "Unknown Error" }, /* 30 */ |
798 | { false, "UNKNOWN" , "Unknown Error" }, /* 31 */ |
799 | }; |
800 | |
801 | static void dwxgmac3_handle_mtl_err(struct net_device *ndev, |
802 | void __iomem *ioaddr, bool correctable, |
803 | struct stmmac_safety_stats *stats) |
804 | { |
805 | u32 value; |
806 | |
807 | value = readl(addr: ioaddr + XGMAC_MTL_ECC_INT_STATUS); |
808 | writel(val: value, addr: ioaddr + XGMAC_MTL_ECC_INT_STATUS); |
809 | |
810 | dwxgmac3_log_error(ndev, value, corr: correctable, module_name: "MTL" , |
811 | desc: dwxgmac3_mtl_errors, STAT_OFF(mtl_errors), stats); |
812 | } |
813 | |
814 | static const struct dwxgmac3_error_desc dwxgmac3_dma_errors[32]= { |
815 | { true, "TCES" , "DMA TSO Memory Error" }, |
816 | { true, "TAMS" , "DMA TSO Memory Address Mismatch Error" }, |
817 | { true, "TUES" , "DMA TSO Memory Error" }, |
818 | { false, "UNKNOWN" , "Unknown Error" }, /* 3 */ |
819 | { true, "DCES" , "DMA DCACHE Memory Error" }, |
820 | { true, "DAMS" , "DMA DCACHE Address Mismatch Error" }, |
821 | { true, "DUES" , "DMA DCACHE Memory Error" }, |
822 | { false, "UNKNOWN" , "Unknown Error" }, /* 7 */ |
823 | { false, "UNKNOWN" , "Unknown Error" }, /* 8 */ |
824 | { false, "UNKNOWN" , "Unknown Error" }, /* 9 */ |
825 | { false, "UNKNOWN" , "Unknown Error" }, /* 10 */ |
826 | { false, "UNKNOWN" , "Unknown Error" }, /* 11 */ |
827 | { false, "UNKNOWN" , "Unknown Error" }, /* 12 */ |
828 | { false, "UNKNOWN" , "Unknown Error" }, /* 13 */ |
829 | { false, "UNKNOWN" , "Unknown Error" }, /* 14 */ |
830 | { false, "UNKNOWN" , "Unknown Error" }, /* 15 */ |
831 | { false, "UNKNOWN" , "Unknown Error" }, /* 16 */ |
832 | { false, "UNKNOWN" , "Unknown Error" }, /* 17 */ |
833 | { false, "UNKNOWN" , "Unknown Error" }, /* 18 */ |
834 | { false, "UNKNOWN" , "Unknown Error" }, /* 19 */ |
835 | { false, "UNKNOWN" , "Unknown Error" }, /* 20 */ |
836 | { false, "UNKNOWN" , "Unknown Error" }, /* 21 */ |
837 | { false, "UNKNOWN" , "Unknown Error" }, /* 22 */ |
838 | { false, "UNKNOWN" , "Unknown Error" }, /* 23 */ |
839 | { false, "UNKNOWN" , "Unknown Error" }, /* 24 */ |
840 | { false, "UNKNOWN" , "Unknown Error" }, /* 25 */ |
841 | { false, "UNKNOWN" , "Unknown Error" }, /* 26 */ |
842 | { false, "UNKNOWN" , "Unknown Error" }, /* 27 */ |
843 | { false, "UNKNOWN" , "Unknown Error" }, /* 28 */ |
844 | { false, "UNKNOWN" , "Unknown Error" }, /* 29 */ |
845 | { false, "UNKNOWN" , "Unknown Error" }, /* 30 */ |
846 | { false, "UNKNOWN" , "Unknown Error" }, /* 31 */ |
847 | }; |
848 | |
849 | #define DPP_RX_ERR "Read Rx Descriptor Parity checker Error" |
850 | #define DPP_TX_ERR "Read Tx Descriptor Parity checker Error" |
851 | |
852 | static const struct dwxgmac3_error_desc dwxgmac3_dma_dpp_errors[32] = { |
853 | { true, "TDPES0" , DPP_TX_ERR }, |
854 | { true, "TDPES1" , DPP_TX_ERR }, |
855 | { true, "TDPES2" , DPP_TX_ERR }, |
856 | { true, "TDPES3" , DPP_TX_ERR }, |
857 | { true, "TDPES4" , DPP_TX_ERR }, |
858 | { true, "TDPES5" , DPP_TX_ERR }, |
859 | { true, "TDPES6" , DPP_TX_ERR }, |
860 | { true, "TDPES7" , DPP_TX_ERR }, |
861 | { true, "TDPES8" , DPP_TX_ERR }, |
862 | { true, "TDPES9" , DPP_TX_ERR }, |
863 | { true, "TDPES10" , DPP_TX_ERR }, |
864 | { true, "TDPES11" , DPP_TX_ERR }, |
865 | { true, "TDPES12" , DPP_TX_ERR }, |
866 | { true, "TDPES13" , DPP_TX_ERR }, |
867 | { true, "TDPES14" , DPP_TX_ERR }, |
868 | { true, "TDPES15" , DPP_TX_ERR }, |
869 | { true, "RDPES0" , DPP_RX_ERR }, |
870 | { true, "RDPES1" , DPP_RX_ERR }, |
871 | { true, "RDPES2" , DPP_RX_ERR }, |
872 | { true, "RDPES3" , DPP_RX_ERR }, |
873 | { true, "RDPES4" , DPP_RX_ERR }, |
874 | { true, "RDPES5" , DPP_RX_ERR }, |
875 | { true, "RDPES6" , DPP_RX_ERR }, |
876 | { true, "RDPES7" , DPP_RX_ERR }, |
877 | { true, "RDPES8" , DPP_RX_ERR }, |
878 | { true, "RDPES9" , DPP_RX_ERR }, |
879 | { true, "RDPES10" , DPP_RX_ERR }, |
880 | { true, "RDPES11" , DPP_RX_ERR }, |
881 | { true, "RDPES12" , DPP_RX_ERR }, |
882 | { true, "RDPES13" , DPP_RX_ERR }, |
883 | { true, "RDPES14" , DPP_RX_ERR }, |
884 | { true, "RDPES15" , DPP_RX_ERR }, |
885 | }; |
886 | |
887 | static void dwxgmac3_handle_dma_err(struct net_device *ndev, |
888 | void __iomem *ioaddr, bool correctable, |
889 | struct stmmac_safety_stats *stats) |
890 | { |
891 | u32 value; |
892 | |
893 | value = readl(addr: ioaddr + XGMAC_DMA_ECC_INT_STATUS); |
894 | writel(val: value, addr: ioaddr + XGMAC_DMA_ECC_INT_STATUS); |
895 | |
896 | dwxgmac3_log_error(ndev, value, corr: correctable, module_name: "DMA" , |
897 | desc: dwxgmac3_dma_errors, STAT_OFF(dma_errors), stats); |
898 | |
899 | value = readl(addr: ioaddr + XGMAC_DMA_DPP_INT_STATUS); |
900 | writel(val: value, addr: ioaddr + XGMAC_DMA_DPP_INT_STATUS); |
901 | |
902 | dwxgmac3_log_error(ndev, value, corr: false, module_name: "DMA_DPP" , |
903 | desc: dwxgmac3_dma_dpp_errors, |
904 | STAT_OFF(dma_dpp_errors), stats); |
905 | } |
906 | |
907 | static int |
908 | dwxgmac3_safety_feat_config(void __iomem *ioaddr, unsigned int asp, |
909 | struct stmmac_safety_feature_cfg *safety_cfg) |
910 | { |
911 | u32 value; |
912 | |
913 | if (!asp) |
914 | return -EINVAL; |
915 | |
916 | /* 1. Enable Safety Features */ |
917 | writel(val: 0x0, addr: ioaddr + XGMAC_MTL_ECC_CONTROL); |
918 | |
919 | /* 2. Enable MTL Safety Interrupts */ |
920 | value = readl(addr: ioaddr + XGMAC_MTL_ECC_INT_ENABLE); |
921 | value |= XGMAC_RPCEIE; /* RX Parser Memory Correctable Error */ |
922 | value |= XGMAC_ECEIE; /* EST Memory Correctable Error */ |
923 | value |= XGMAC_RXCEIE; /* RX Memory Correctable Error */ |
924 | value |= XGMAC_TXCEIE; /* TX Memory Correctable Error */ |
925 | writel(val: value, addr: ioaddr + XGMAC_MTL_ECC_INT_ENABLE); |
926 | |
927 | /* 3. Enable DMA Safety Interrupts */ |
928 | value = readl(addr: ioaddr + XGMAC_DMA_ECC_INT_ENABLE); |
929 | value |= XGMAC_DCEIE; /* Descriptor Cache Memory Correctable Error */ |
930 | value |= XGMAC_TCEIE; /* TSO Memory Correctable Error */ |
931 | writel(val: value, addr: ioaddr + XGMAC_DMA_ECC_INT_ENABLE); |
932 | |
933 | /* 0x2: Without ECC or Parity Ports on External Application Interface |
934 | * 0x4: Only ECC Protection for External Memory feature is selected |
935 | */ |
936 | if (asp == 0x2 || asp == 0x4) |
937 | return 0; |
938 | |
939 | /* 4. Enable Parity and Timeout for FSM */ |
940 | value = readl(addr: ioaddr + XGMAC_MAC_FSM_CONTROL); |
941 | value |= XGMAC_PRTYEN; /* FSM Parity Feature */ |
942 | value |= XGMAC_TMOUTEN; /* FSM Timeout Feature */ |
943 | writel(val: value, addr: ioaddr + XGMAC_MAC_FSM_CONTROL); |
944 | |
945 | /* 5. Enable Data Path Parity Protection */ |
946 | value = readl(addr: ioaddr + XGMAC_MTL_DPP_CONTROL); |
947 | /* already enabled by default, explicit enable it again */ |
948 | value &= ~XGMAC_DPP_DISABLE; |
949 | writel(val: value, addr: ioaddr + XGMAC_MTL_DPP_CONTROL); |
950 | |
951 | return 0; |
952 | } |
953 | |
954 | static int dwxgmac3_safety_feat_irq_status(struct net_device *ndev, |
955 | void __iomem *ioaddr, |
956 | unsigned int asp, |
957 | struct stmmac_safety_stats *stats) |
958 | { |
959 | bool err, corr; |
960 | u32 mtl, dma; |
961 | int ret = 0; |
962 | |
963 | if (!asp) |
964 | return -EINVAL; |
965 | |
966 | mtl = readl(addr: ioaddr + XGMAC_MTL_SAFETY_INT_STATUS); |
967 | dma = readl(addr: ioaddr + XGMAC_DMA_SAFETY_INT_STATUS); |
968 | |
969 | err = (mtl & XGMAC_MCSIS) || (dma & XGMAC_MCSIS); |
970 | corr = false; |
971 | if (err) { |
972 | dwxgmac3_handle_mac_err(ndev, ioaddr, correctable: corr, stats); |
973 | ret |= !corr; |
974 | } |
975 | |
976 | err = (mtl & (XGMAC_MEUIS | XGMAC_MECIS)) || |
977 | (dma & (XGMAC_MSUIS | XGMAC_MSCIS)); |
978 | corr = (mtl & XGMAC_MECIS) || (dma & XGMAC_MSCIS); |
979 | if (err) { |
980 | dwxgmac3_handle_mtl_err(ndev, ioaddr, correctable: corr, stats); |
981 | ret |= !corr; |
982 | } |
983 | |
984 | /* DMA_DPP_Interrupt_Status is indicated by MCSIS bit in |
985 | * DMA_Safety_Interrupt_Status, so we handle DMA Data Path |
986 | * Parity Errors here |
987 | */ |
988 | err = dma & (XGMAC_DEUIS | XGMAC_DECIS | XGMAC_MCSIS); |
989 | corr = dma & XGMAC_DECIS; |
990 | if (err) { |
991 | dwxgmac3_handle_dma_err(ndev, ioaddr, correctable: corr, stats); |
992 | ret |= !corr; |
993 | } |
994 | |
995 | return ret; |
996 | } |
997 | |
998 | static const struct dwxgmac3_error { |
999 | const struct dwxgmac3_error_desc *desc; |
1000 | } dwxgmac3_all_errors[] = { |
1001 | { dwxgmac3_mac_errors }, |
1002 | { dwxgmac3_mtl_errors }, |
1003 | { dwxgmac3_dma_errors }, |
1004 | { dwxgmac3_dma_dpp_errors }, |
1005 | }; |
1006 | |
1007 | static int dwxgmac3_safety_feat_dump(struct stmmac_safety_stats *stats, |
1008 | int index, unsigned long *count, |
1009 | const char **desc) |
1010 | { |
1011 | int module = index / 32, offset = index % 32; |
1012 | unsigned long *ptr = (unsigned long *)stats; |
1013 | |
1014 | if (module >= ARRAY_SIZE(dwxgmac3_all_errors)) |
1015 | return -EINVAL; |
1016 | if (!dwxgmac3_all_errors[module].desc[offset].valid) |
1017 | return -EINVAL; |
1018 | if (count) |
1019 | *count = *(ptr + index); |
1020 | if (desc) |
1021 | *desc = dwxgmac3_all_errors[module].desc[offset].desc; |
1022 | return 0; |
1023 | } |
1024 | |
1025 | static int dwxgmac3_rxp_disable(void __iomem *ioaddr) |
1026 | { |
1027 | u32 val = readl(addr: ioaddr + XGMAC_MTL_OPMODE); |
1028 | |
1029 | val &= ~XGMAC_FRPE; |
1030 | writel(val, addr: ioaddr + XGMAC_MTL_OPMODE); |
1031 | |
1032 | return 0; |
1033 | } |
1034 | |
1035 | static void dwxgmac3_rxp_enable(void __iomem *ioaddr) |
1036 | { |
1037 | u32 val; |
1038 | |
1039 | val = readl(addr: ioaddr + XGMAC_MTL_OPMODE); |
1040 | val |= XGMAC_FRPE; |
1041 | writel(val, addr: ioaddr + XGMAC_MTL_OPMODE); |
1042 | } |
1043 | |
1044 | static int dwxgmac3_rxp_update_single_entry(void __iomem *ioaddr, |
1045 | struct stmmac_tc_entry *entry, |
1046 | int pos) |
1047 | { |
1048 | int ret, i; |
1049 | |
1050 | for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) { |
1051 | int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i; |
1052 | u32 val; |
1053 | |
1054 | /* Wait for ready */ |
1055 | ret = readl_poll_timeout(ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST, |
1056 | val, !(val & XGMAC_STARTBUSY), 1, 10000); |
1057 | if (ret) |
1058 | return ret; |
1059 | |
1060 | /* Write data */ |
1061 | val = *((u32 *)&entry->val + i); |
1062 | writel(val, addr: ioaddr + XGMAC_MTL_RXP_IACC_DATA); |
1063 | |
1064 | /* Write pos */ |
1065 | val = real_pos & XGMAC_ADDR; |
1066 | writel(val, addr: ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST); |
1067 | |
1068 | /* Write OP */ |
1069 | val |= XGMAC_WRRDN; |
1070 | writel(val, addr: ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST); |
1071 | |
1072 | /* Start Write */ |
1073 | val |= XGMAC_STARTBUSY; |
1074 | writel(val, addr: ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST); |
1075 | |
1076 | /* Wait for done */ |
1077 | ret = readl_poll_timeout(ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST, |
1078 | val, !(val & XGMAC_STARTBUSY), 1, 10000); |
1079 | if (ret) |
1080 | return ret; |
1081 | } |
1082 | |
1083 | return 0; |
1084 | } |
1085 | |
1086 | static struct stmmac_tc_entry * |
1087 | dwxgmac3_rxp_get_next_entry(struct stmmac_tc_entry *entries, |
1088 | unsigned int count, u32 curr_prio) |
1089 | { |
1090 | struct stmmac_tc_entry *entry; |
1091 | u32 min_prio = ~0x0; |
1092 | int i, min_prio_idx; |
1093 | bool found = false; |
1094 | |
1095 | for (i = count - 1; i >= 0; i--) { |
1096 | entry = &entries[i]; |
1097 | |
1098 | /* Do not update unused entries */ |
1099 | if (!entry->in_use) |
1100 | continue; |
1101 | /* Do not update already updated entries (i.e. fragments) */ |
1102 | if (entry->in_hw) |
1103 | continue; |
1104 | /* Let last entry be updated last */ |
1105 | if (entry->is_last) |
1106 | continue; |
1107 | /* Do not return fragments */ |
1108 | if (entry->is_frag) |
1109 | continue; |
1110 | /* Check if we already checked this prio */ |
1111 | if (entry->prio < curr_prio) |
1112 | continue; |
1113 | /* Check if this is the minimum prio */ |
1114 | if (entry->prio < min_prio) { |
1115 | min_prio = entry->prio; |
1116 | min_prio_idx = i; |
1117 | found = true; |
1118 | } |
1119 | } |
1120 | |
1121 | if (found) |
1122 | return &entries[min_prio_idx]; |
1123 | return NULL; |
1124 | } |
1125 | |
1126 | static int dwxgmac3_rxp_config(void __iomem *ioaddr, |
1127 | struct stmmac_tc_entry *entries, |
1128 | unsigned int count) |
1129 | { |
1130 | struct stmmac_tc_entry *entry, *frag; |
1131 | int i, ret, nve = 0; |
1132 | u32 curr_prio = 0; |
1133 | u32 old_val, val; |
1134 | |
1135 | /* Force disable RX */ |
1136 | old_val = readl(addr: ioaddr + XGMAC_RX_CONFIG); |
1137 | val = old_val & ~XGMAC_CONFIG_RE; |
1138 | writel(val, addr: ioaddr + XGMAC_RX_CONFIG); |
1139 | |
1140 | /* Disable RX Parser */ |
1141 | ret = dwxgmac3_rxp_disable(ioaddr); |
1142 | if (ret) |
1143 | goto re_enable; |
1144 | |
1145 | /* Set all entries as NOT in HW */ |
1146 | for (i = 0; i < count; i++) { |
1147 | entry = &entries[i]; |
1148 | entry->in_hw = false; |
1149 | } |
1150 | |
1151 | /* Update entries by reverse order */ |
1152 | while (1) { |
1153 | entry = dwxgmac3_rxp_get_next_entry(entries, count, curr_prio); |
1154 | if (!entry) |
1155 | break; |
1156 | |
1157 | curr_prio = entry->prio; |
1158 | frag = entry->frag_ptr; |
1159 | |
1160 | /* Set special fragment requirements */ |
1161 | if (frag) { |
1162 | entry->val.af = 0; |
1163 | entry->val.rf = 0; |
1164 | entry->val.nc = 1; |
1165 | entry->val.ok_index = nve + 2; |
1166 | } |
1167 | |
1168 | ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, pos: nve); |
1169 | if (ret) |
1170 | goto re_enable; |
1171 | |
1172 | entry->table_pos = nve++; |
1173 | entry->in_hw = true; |
1174 | |
1175 | if (frag && !frag->in_hw) { |
1176 | ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry: frag, pos: nve); |
1177 | if (ret) |
1178 | goto re_enable; |
1179 | frag->table_pos = nve++; |
1180 | frag->in_hw = true; |
1181 | } |
1182 | } |
1183 | |
1184 | if (!nve) |
1185 | goto re_enable; |
1186 | |
1187 | /* Update all pass entry */ |
1188 | for (i = 0; i < count; i++) { |
1189 | entry = &entries[i]; |
1190 | if (!entry->is_last) |
1191 | continue; |
1192 | |
1193 | ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, pos: nve); |
1194 | if (ret) |
1195 | goto re_enable; |
1196 | |
1197 | entry->table_pos = nve++; |
1198 | } |
1199 | |
1200 | /* Assume n. of parsable entries == n. of valid entries */ |
1201 | val = (nve << 16) & XGMAC_NPE; |
1202 | val |= nve & XGMAC_NVE; |
1203 | writel(val, addr: ioaddr + XGMAC_MTL_RXP_CONTROL_STATUS); |
1204 | |
1205 | /* Enable RX Parser */ |
1206 | dwxgmac3_rxp_enable(ioaddr); |
1207 | |
1208 | re_enable: |
1209 | /* Re-enable RX */ |
1210 | writel(val: old_val, addr: ioaddr + XGMAC_RX_CONFIG); |
1211 | return ret; |
1212 | } |
1213 | |
1214 | static int dwxgmac2_get_mac_tx_timestamp(struct mac_device_info *hw, u64 *ts) |
1215 | { |
1216 | void __iomem *ioaddr = hw->pcsr; |
1217 | u32 value; |
1218 | |
1219 | if (readl_poll_timeout_atomic(ioaddr + XGMAC_TIMESTAMP_STATUS, |
1220 | value, value & XGMAC_TXTSC, 100, 10000)) |
1221 | return -EBUSY; |
1222 | |
1223 | *ts = readl(addr: ioaddr + XGMAC_TXTIMESTAMP_NSEC) & XGMAC_TXTSSTSLO; |
1224 | *ts += readl(addr: ioaddr + XGMAC_TXTIMESTAMP_SEC) * 1000000000ULL; |
1225 | return 0; |
1226 | } |
1227 | |
1228 | static int dwxgmac2_flex_pps_config(void __iomem *ioaddr, int index, |
1229 | struct stmmac_pps_cfg *cfg, bool enable, |
1230 | u32 sub_second_inc, u32 systime_flags) |
1231 | { |
1232 | u32 tnsec = readl(addr: ioaddr + XGMAC_PPSx_TARGET_TIME_NSEC(index)); |
1233 | u32 val = readl(addr: ioaddr + XGMAC_PPS_CONTROL); |
1234 | u64 period; |
1235 | |
1236 | if (!cfg->available) |
1237 | return -EINVAL; |
1238 | if (tnsec & XGMAC_TRGTBUSY0) |
1239 | return -EBUSY; |
1240 | if (!sub_second_inc || !systime_flags) |
1241 | return -EINVAL; |
1242 | |
1243 | val &= ~XGMAC_PPSx_MASK(index); |
1244 | |
1245 | if (!enable) { |
1246 | val |= XGMAC_PPSCMDx(index, XGMAC_PPSCMD_STOP); |
1247 | writel(val, addr: ioaddr + XGMAC_PPS_CONTROL); |
1248 | return 0; |
1249 | } |
1250 | |
1251 | val |= XGMAC_PPSCMDx(index, XGMAC_PPSCMD_START); |
1252 | val |= XGMAC_TRGTMODSELx(index, XGMAC_PPSCMD_START); |
1253 | |
1254 | /* XGMAC Core has 4 PPS outputs at most. |
1255 | * |
1256 | * Prior XGMAC Core 3.20, Fixed mode or Flexible mode are selectable for |
1257 | * PPS0 only via PPSEN0. PPS{1,2,3} are in Flexible mode by default, |
1258 | * and can not be switched to Fixed mode, since PPSEN{1,2,3} are |
1259 | * read-only reserved to 0. |
1260 | * But we always set PPSEN{1,2,3} do not make things worse ;-) |
1261 | * |
1262 | * From XGMAC Core 3.20 and later, PPSEN{0,1,2,3} are writable and must |
1263 | * be set, or the PPS outputs stay in Fixed PPS mode by default. |
1264 | */ |
1265 | val |= XGMAC_PPSENx(index); |
1266 | |
1267 | writel(val: cfg->start.tv_sec, addr: ioaddr + XGMAC_PPSx_TARGET_TIME_SEC(index)); |
1268 | |
1269 | if (!(systime_flags & PTP_TCR_TSCTRLSSR)) |
1270 | cfg->start.tv_nsec = (cfg->start.tv_nsec * 1000) / 465; |
1271 | writel(val: cfg->start.tv_nsec, addr: ioaddr + XGMAC_PPSx_TARGET_TIME_NSEC(index)); |
1272 | |
1273 | period = cfg->period.tv_sec * 1000000000; |
1274 | period += cfg->period.tv_nsec; |
1275 | |
1276 | do_div(period, sub_second_inc); |
1277 | |
1278 | if (period <= 1) |
1279 | return -EINVAL; |
1280 | |
1281 | writel(val: period - 1, addr: ioaddr + XGMAC_PPSx_INTERVAL(index)); |
1282 | |
1283 | period >>= 1; |
1284 | if (period <= 1) |
1285 | return -EINVAL; |
1286 | |
1287 | writel(val: period - 1, addr: ioaddr + XGMAC_PPSx_WIDTH(index)); |
1288 | |
1289 | /* Finally, activate it */ |
1290 | writel(val, addr: ioaddr + XGMAC_PPS_CONTROL); |
1291 | return 0; |
1292 | } |
1293 | |
1294 | static void dwxgmac2_sarc_configure(void __iomem *ioaddr, int val) |
1295 | { |
1296 | u32 value = readl(addr: ioaddr + XGMAC_TX_CONFIG); |
1297 | |
1298 | value &= ~XGMAC_CONFIG_SARC; |
1299 | value |= val << XGMAC_CONFIG_SARC_SHIFT; |
1300 | |
1301 | writel(val: value, addr: ioaddr + XGMAC_TX_CONFIG); |
1302 | } |
1303 | |
1304 | static void dwxgmac2_enable_vlan(struct mac_device_info *hw, u32 type) |
1305 | { |
1306 | void __iomem *ioaddr = hw->pcsr; |
1307 | u32 value; |
1308 | |
1309 | value = readl(addr: ioaddr + XGMAC_VLAN_INCL); |
1310 | value |= XGMAC_VLAN_VLTI; |
1311 | value |= XGMAC_VLAN_CSVL; /* Only use SVLAN */ |
1312 | value &= ~XGMAC_VLAN_VLC; |
1313 | value |= (type << XGMAC_VLAN_VLC_SHIFT) & XGMAC_VLAN_VLC; |
1314 | writel(val: value, addr: ioaddr + XGMAC_VLAN_INCL); |
1315 | } |
1316 | |
1317 | static int dwxgmac2_filter_wait(struct mac_device_info *hw) |
1318 | { |
1319 | void __iomem *ioaddr = hw->pcsr; |
1320 | u32 value; |
1321 | |
1322 | if (readl_poll_timeout(ioaddr + XGMAC_L3L4_ADDR_CTRL, value, |
1323 | !(value & XGMAC_XB), 100, 10000)) |
1324 | return -EBUSY; |
1325 | return 0; |
1326 | } |
1327 | |
1328 | static int dwxgmac2_filter_read(struct mac_device_info *hw, u32 filter_no, |
1329 | u8 reg, u32 *data) |
1330 | { |
1331 | void __iomem *ioaddr = hw->pcsr; |
1332 | u32 value; |
1333 | int ret; |
1334 | |
1335 | ret = dwxgmac2_filter_wait(hw); |
1336 | if (ret) |
1337 | return ret; |
1338 | |
1339 | value = ((filter_no << XGMAC_IDDR_FNUM) | reg) << XGMAC_IDDR_SHIFT; |
1340 | value |= XGMAC_TT | XGMAC_XB; |
1341 | writel(val: value, addr: ioaddr + XGMAC_L3L4_ADDR_CTRL); |
1342 | |
1343 | ret = dwxgmac2_filter_wait(hw); |
1344 | if (ret) |
1345 | return ret; |
1346 | |
1347 | *data = readl(addr: ioaddr + XGMAC_L3L4_DATA); |
1348 | return 0; |
1349 | } |
1350 | |
1351 | static int dwxgmac2_filter_write(struct mac_device_info *hw, u32 filter_no, |
1352 | u8 reg, u32 data) |
1353 | { |
1354 | void __iomem *ioaddr = hw->pcsr; |
1355 | u32 value; |
1356 | int ret; |
1357 | |
1358 | ret = dwxgmac2_filter_wait(hw); |
1359 | if (ret) |
1360 | return ret; |
1361 | |
1362 | writel(val: data, addr: ioaddr + XGMAC_L3L4_DATA); |
1363 | |
1364 | value = ((filter_no << XGMAC_IDDR_FNUM) | reg) << XGMAC_IDDR_SHIFT; |
1365 | value |= XGMAC_XB; |
1366 | writel(val: value, addr: ioaddr + XGMAC_L3L4_ADDR_CTRL); |
1367 | |
1368 | return dwxgmac2_filter_wait(hw); |
1369 | } |
1370 | |
1371 | static int dwxgmac2_config_l3_filter(struct mac_device_info *hw, u32 filter_no, |
1372 | bool en, bool ipv6, bool sa, bool inv, |
1373 | u32 match) |
1374 | { |
1375 | void __iomem *ioaddr = hw->pcsr; |
1376 | u32 value; |
1377 | int ret; |
1378 | |
1379 | value = readl(addr: ioaddr + XGMAC_PACKET_FILTER); |
1380 | value |= XGMAC_FILTER_IPFE; |
1381 | writel(val: value, addr: ioaddr + XGMAC_PACKET_FILTER); |
1382 | |
1383 | ret = dwxgmac2_filter_read(hw, filter_no, XGMAC_L3L4_CTRL, data: &value); |
1384 | if (ret) |
1385 | return ret; |
1386 | |
1387 | /* For IPv6 not both SA/DA filters can be active */ |
1388 | if (ipv6) { |
1389 | value |= XGMAC_L3PEN0; |
1390 | value &= ~(XGMAC_L3SAM0 | XGMAC_L3SAIM0); |
1391 | value &= ~(XGMAC_L3DAM0 | XGMAC_L3DAIM0); |
1392 | if (sa) { |
1393 | value |= XGMAC_L3SAM0; |
1394 | if (inv) |
1395 | value |= XGMAC_L3SAIM0; |
1396 | } else { |
1397 | value |= XGMAC_L3DAM0; |
1398 | if (inv) |
1399 | value |= XGMAC_L3DAIM0; |
1400 | } |
1401 | } else { |
1402 | value &= ~XGMAC_L3PEN0; |
1403 | if (sa) { |
1404 | value |= XGMAC_L3SAM0; |
1405 | if (inv) |
1406 | value |= XGMAC_L3SAIM0; |
1407 | } else { |
1408 | value |= XGMAC_L3DAM0; |
1409 | if (inv) |
1410 | value |= XGMAC_L3DAIM0; |
1411 | } |
1412 | } |
1413 | |
1414 | ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, data: value); |
1415 | if (ret) |
1416 | return ret; |
1417 | |
1418 | if (sa) { |
1419 | ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3_ADDR0, data: match); |
1420 | if (ret) |
1421 | return ret; |
1422 | } else { |
1423 | ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3_ADDR1, data: match); |
1424 | if (ret) |
1425 | return ret; |
1426 | } |
1427 | |
1428 | if (!en) |
1429 | return dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, data: 0); |
1430 | |
1431 | return 0; |
1432 | } |
1433 | |
1434 | static int dwxgmac2_config_l4_filter(struct mac_device_info *hw, u32 filter_no, |
1435 | bool en, bool udp, bool sa, bool inv, |
1436 | u32 match) |
1437 | { |
1438 | void __iomem *ioaddr = hw->pcsr; |
1439 | u32 value; |
1440 | int ret; |
1441 | |
1442 | value = readl(addr: ioaddr + XGMAC_PACKET_FILTER); |
1443 | value |= XGMAC_FILTER_IPFE; |
1444 | writel(val: value, addr: ioaddr + XGMAC_PACKET_FILTER); |
1445 | |
1446 | ret = dwxgmac2_filter_read(hw, filter_no, XGMAC_L3L4_CTRL, data: &value); |
1447 | if (ret) |
1448 | return ret; |
1449 | |
1450 | if (udp) { |
1451 | value |= XGMAC_L4PEN0; |
1452 | } else { |
1453 | value &= ~XGMAC_L4PEN0; |
1454 | } |
1455 | |
1456 | value &= ~(XGMAC_L4SPM0 | XGMAC_L4SPIM0); |
1457 | value &= ~(XGMAC_L4DPM0 | XGMAC_L4DPIM0); |
1458 | if (sa) { |
1459 | value |= XGMAC_L4SPM0; |
1460 | if (inv) |
1461 | value |= XGMAC_L4SPIM0; |
1462 | } else { |
1463 | value |= XGMAC_L4DPM0; |
1464 | if (inv) |
1465 | value |= XGMAC_L4DPIM0; |
1466 | } |
1467 | |
1468 | ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, data: value); |
1469 | if (ret) |
1470 | return ret; |
1471 | |
1472 | if (sa) { |
1473 | value = match & XGMAC_L4SP0; |
1474 | |
1475 | ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L4_ADDR, data: value); |
1476 | if (ret) |
1477 | return ret; |
1478 | } else { |
1479 | value = (match << XGMAC_L4DP0_SHIFT) & XGMAC_L4DP0; |
1480 | |
1481 | ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L4_ADDR, data: value); |
1482 | if (ret) |
1483 | return ret; |
1484 | } |
1485 | |
1486 | if (!en) |
1487 | return dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, data: 0); |
1488 | |
1489 | return 0; |
1490 | } |
1491 | |
1492 | static void dwxgmac2_set_arp_offload(struct mac_device_info *hw, bool en, |
1493 | u32 addr) |
1494 | { |
1495 | void __iomem *ioaddr = hw->pcsr; |
1496 | u32 value; |
1497 | |
1498 | writel(val: addr, addr: ioaddr + XGMAC_ARP_ADDR); |
1499 | |
1500 | value = readl(addr: ioaddr + XGMAC_RX_CONFIG); |
1501 | if (en) |
1502 | value |= XGMAC_CONFIG_ARPEN; |
1503 | else |
1504 | value &= ~XGMAC_CONFIG_ARPEN; |
1505 | writel(val: value, addr: ioaddr + XGMAC_RX_CONFIG); |
1506 | } |
1507 | |
1508 | static void dwxgmac3_fpe_configure(void __iomem *ioaddr, struct stmmac_fpe_cfg *cfg, |
1509 | u32 num_txq, |
1510 | u32 num_rxq, bool enable) |
1511 | { |
1512 | u32 value; |
1513 | |
1514 | if (!enable) { |
1515 | value = readl(addr: ioaddr + XGMAC_FPE_CTRL_STS); |
1516 | |
1517 | value &= ~XGMAC_EFPE; |
1518 | |
1519 | writel(val: value, addr: ioaddr + XGMAC_FPE_CTRL_STS); |
1520 | return; |
1521 | } |
1522 | |
1523 | value = readl(addr: ioaddr + XGMAC_RXQ_CTRL1); |
1524 | value &= ~XGMAC_RQ; |
1525 | value |= (num_rxq - 1) << XGMAC_RQ_SHIFT; |
1526 | writel(val: value, addr: ioaddr + XGMAC_RXQ_CTRL1); |
1527 | |
1528 | value = readl(addr: ioaddr + XGMAC_FPE_CTRL_STS); |
1529 | value |= XGMAC_EFPE; |
1530 | writel(val: value, addr: ioaddr + XGMAC_FPE_CTRL_STS); |
1531 | } |
1532 | |
1533 | const struct stmmac_ops dwxgmac210_ops = { |
1534 | .core_init = dwxgmac2_core_init, |
1535 | .set_mac = dwxgmac2_set_mac, |
1536 | .rx_ipc = dwxgmac2_rx_ipc, |
1537 | .rx_queue_enable = dwxgmac2_rx_queue_enable, |
1538 | .rx_queue_prio = dwxgmac2_rx_queue_prio, |
1539 | .tx_queue_prio = dwxgmac2_tx_queue_prio, |
1540 | .rx_queue_routing = dwxgmac2_rx_queue_routing, |
1541 | .prog_mtl_rx_algorithms = dwxgmac2_prog_mtl_rx_algorithms, |
1542 | .prog_mtl_tx_algorithms = dwxgmac2_prog_mtl_tx_algorithms, |
1543 | .set_mtl_tx_queue_weight = dwxgmac2_set_mtl_tx_queue_weight, |
1544 | .map_mtl_to_dma = dwxgmac2_map_mtl_to_dma, |
1545 | .config_cbs = dwxgmac2_config_cbs, |
1546 | .dump_regs = dwxgmac2_dump_regs, |
1547 | .host_irq_status = dwxgmac2_host_irq_status, |
1548 | .host_mtl_irq_status = dwxgmac2_host_mtl_irq_status, |
1549 | .flow_ctrl = dwxgmac2_flow_ctrl, |
1550 | .pmt = dwxgmac2_pmt, |
1551 | .set_umac_addr = dwxgmac2_set_umac_addr, |
1552 | .get_umac_addr = dwxgmac2_get_umac_addr, |
1553 | .set_eee_mode = dwxgmac2_set_eee_mode, |
1554 | .reset_eee_mode = dwxgmac2_reset_eee_mode, |
1555 | .set_eee_timer = dwxgmac2_set_eee_timer, |
1556 | .set_eee_pls = dwxgmac2_set_eee_pls, |
1557 | .pcs_ctrl_ane = NULL, |
1558 | .pcs_rane = NULL, |
1559 | .pcs_get_adv_lp = NULL, |
1560 | .debug = NULL, |
1561 | .set_filter = dwxgmac2_set_filter, |
1562 | .safety_feat_config = dwxgmac3_safety_feat_config, |
1563 | .safety_feat_irq_status = dwxgmac3_safety_feat_irq_status, |
1564 | .safety_feat_dump = dwxgmac3_safety_feat_dump, |
1565 | .set_mac_loopback = dwxgmac2_set_mac_loopback, |
1566 | .rss_configure = dwxgmac2_rss_configure, |
1567 | .update_vlan_hash = dwxgmac2_update_vlan_hash, |
1568 | .rxp_config = dwxgmac3_rxp_config, |
1569 | .get_mac_tx_timestamp = dwxgmac2_get_mac_tx_timestamp, |
1570 | .flex_pps_config = dwxgmac2_flex_pps_config, |
1571 | .sarc_configure = dwxgmac2_sarc_configure, |
1572 | .enable_vlan = dwxgmac2_enable_vlan, |
1573 | .config_l3_filter = dwxgmac2_config_l3_filter, |
1574 | .config_l4_filter = dwxgmac2_config_l4_filter, |
1575 | .set_arp_offload = dwxgmac2_set_arp_offload, |
1576 | .fpe_configure = dwxgmac3_fpe_configure, |
1577 | }; |
1578 | |
1579 | static void dwxlgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode, |
1580 | u32 queue) |
1581 | { |
1582 | void __iomem *ioaddr = hw->pcsr; |
1583 | u32 value; |
1584 | |
1585 | value = readl(addr: ioaddr + XLGMAC_RXQ_ENABLE_CTRL0) & ~XGMAC_RXQEN(queue); |
1586 | if (mode == MTL_QUEUE_AVB) |
1587 | value |= 0x1 << XGMAC_RXQEN_SHIFT(queue); |
1588 | else if (mode == MTL_QUEUE_DCB) |
1589 | value |= 0x2 << XGMAC_RXQEN_SHIFT(queue); |
1590 | writel(val: value, addr: ioaddr + XLGMAC_RXQ_ENABLE_CTRL0); |
1591 | } |
1592 | |
1593 | const struct stmmac_ops dwxlgmac2_ops = { |
1594 | .core_init = dwxgmac2_core_init, |
1595 | .set_mac = dwxgmac2_set_mac, |
1596 | .rx_ipc = dwxgmac2_rx_ipc, |
1597 | .rx_queue_enable = dwxlgmac2_rx_queue_enable, |
1598 | .rx_queue_prio = dwxgmac2_rx_queue_prio, |
1599 | .tx_queue_prio = dwxgmac2_tx_queue_prio, |
1600 | .rx_queue_routing = dwxgmac2_rx_queue_routing, |
1601 | .prog_mtl_rx_algorithms = dwxgmac2_prog_mtl_rx_algorithms, |
1602 | .prog_mtl_tx_algorithms = dwxgmac2_prog_mtl_tx_algorithms, |
1603 | .set_mtl_tx_queue_weight = dwxgmac2_set_mtl_tx_queue_weight, |
1604 | .map_mtl_to_dma = dwxgmac2_map_mtl_to_dma, |
1605 | .config_cbs = dwxgmac2_config_cbs, |
1606 | .dump_regs = dwxgmac2_dump_regs, |
1607 | .host_irq_status = dwxgmac2_host_irq_status, |
1608 | .host_mtl_irq_status = dwxgmac2_host_mtl_irq_status, |
1609 | .flow_ctrl = dwxgmac2_flow_ctrl, |
1610 | .pmt = dwxgmac2_pmt, |
1611 | .set_umac_addr = dwxgmac2_set_umac_addr, |
1612 | .get_umac_addr = dwxgmac2_get_umac_addr, |
1613 | .set_eee_mode = dwxgmac2_set_eee_mode, |
1614 | .reset_eee_mode = dwxgmac2_reset_eee_mode, |
1615 | .set_eee_timer = dwxgmac2_set_eee_timer, |
1616 | .set_eee_pls = dwxgmac2_set_eee_pls, |
1617 | .pcs_ctrl_ane = NULL, |
1618 | .pcs_rane = NULL, |
1619 | .pcs_get_adv_lp = NULL, |
1620 | .debug = NULL, |
1621 | .set_filter = dwxgmac2_set_filter, |
1622 | .safety_feat_config = dwxgmac3_safety_feat_config, |
1623 | .safety_feat_irq_status = dwxgmac3_safety_feat_irq_status, |
1624 | .safety_feat_dump = dwxgmac3_safety_feat_dump, |
1625 | .set_mac_loopback = dwxgmac2_set_mac_loopback, |
1626 | .rss_configure = dwxgmac2_rss_configure, |
1627 | .update_vlan_hash = dwxgmac2_update_vlan_hash, |
1628 | .rxp_config = dwxgmac3_rxp_config, |
1629 | .get_mac_tx_timestamp = dwxgmac2_get_mac_tx_timestamp, |
1630 | .flex_pps_config = dwxgmac2_flex_pps_config, |
1631 | .sarc_configure = dwxgmac2_sarc_configure, |
1632 | .enable_vlan = dwxgmac2_enable_vlan, |
1633 | .config_l3_filter = dwxgmac2_config_l3_filter, |
1634 | .config_l4_filter = dwxgmac2_config_l4_filter, |
1635 | .set_arp_offload = dwxgmac2_set_arp_offload, |
1636 | .fpe_configure = dwxgmac3_fpe_configure, |
1637 | }; |
1638 | |
1639 | int dwxgmac2_setup(struct stmmac_priv *priv) |
1640 | { |
1641 | struct mac_device_info *mac = priv->hw; |
1642 | |
1643 | dev_info(priv->device, "\tXGMAC2\n" ); |
1644 | |
1645 | priv->dev->priv_flags |= IFF_UNICAST_FLT; |
1646 | mac->pcsr = priv->ioaddr; |
1647 | mac->multicast_filter_bins = priv->plat->multicast_filter_bins; |
1648 | mac->unicast_filter_entries = priv->plat->unicast_filter_entries; |
1649 | mac->mcast_bits_log2 = 0; |
1650 | |
1651 | if (mac->multicast_filter_bins) |
1652 | mac->mcast_bits_log2 = ilog2(mac->multicast_filter_bins); |
1653 | |
1654 | mac->link.caps = MAC_ASYM_PAUSE | MAC_SYM_PAUSE | |
1655 | MAC_1000FD | MAC_2500FD | MAC_5000FD | |
1656 | MAC_10000FD; |
1657 | mac->link.duplex = 0; |
1658 | mac->link.speed10 = XGMAC_CONFIG_SS_10_MII; |
1659 | mac->link.speed100 = XGMAC_CONFIG_SS_100_MII; |
1660 | mac->link.speed1000 = XGMAC_CONFIG_SS_1000_GMII; |
1661 | mac->link.speed2500 = XGMAC_CONFIG_SS_2500_GMII; |
1662 | mac->link.xgmii.speed2500 = XGMAC_CONFIG_SS_2500; |
1663 | mac->link.xgmii.speed5000 = XGMAC_CONFIG_SS_5000; |
1664 | mac->link.xgmii.speed10000 = XGMAC_CONFIG_SS_10000; |
1665 | mac->link.speed_mask = XGMAC_CONFIG_SS_MASK; |
1666 | |
1667 | mac->mii.addr = XGMAC_MDIO_ADDR; |
1668 | mac->mii.data = XGMAC_MDIO_DATA; |
1669 | mac->mii.addr_shift = 16; |
1670 | mac->mii.addr_mask = GENMASK(20, 16); |
1671 | mac->mii.reg_shift = 0; |
1672 | mac->mii.reg_mask = GENMASK(15, 0); |
1673 | mac->mii.clk_csr_shift = 19; |
1674 | mac->mii.clk_csr_mask = GENMASK(21, 19); |
1675 | |
1676 | return 0; |
1677 | } |
1678 | |
1679 | int dwxlgmac2_setup(struct stmmac_priv *priv) |
1680 | { |
1681 | struct mac_device_info *mac = priv->hw; |
1682 | |
1683 | dev_info(priv->device, "\tXLGMAC\n" ); |
1684 | |
1685 | priv->dev->priv_flags |= IFF_UNICAST_FLT; |
1686 | mac->pcsr = priv->ioaddr; |
1687 | mac->multicast_filter_bins = priv->plat->multicast_filter_bins; |
1688 | mac->unicast_filter_entries = priv->plat->unicast_filter_entries; |
1689 | mac->mcast_bits_log2 = 0; |
1690 | |
1691 | if (mac->multicast_filter_bins) |
1692 | mac->mcast_bits_log2 = ilog2(mac->multicast_filter_bins); |
1693 | |
1694 | mac->link.caps = MAC_ASYM_PAUSE | MAC_SYM_PAUSE | |
1695 | MAC_1000FD | MAC_2500FD | MAC_5000FD | |
1696 | MAC_10000FD | MAC_25000FD | |
1697 | MAC_40000FD | MAC_50000FD | |
1698 | MAC_100000FD; |
1699 | mac->link.duplex = 0; |
1700 | mac->link.speed1000 = XLGMAC_CONFIG_SS_1000; |
1701 | mac->link.speed2500 = XLGMAC_CONFIG_SS_2500; |
1702 | mac->link.xgmii.speed10000 = XLGMAC_CONFIG_SS_10G; |
1703 | mac->link.xlgmii.speed25000 = XLGMAC_CONFIG_SS_25G; |
1704 | mac->link.xlgmii.speed40000 = XLGMAC_CONFIG_SS_40G; |
1705 | mac->link.xlgmii.speed50000 = XLGMAC_CONFIG_SS_50G; |
1706 | mac->link.xlgmii.speed100000 = XLGMAC_CONFIG_SS_100G; |
1707 | mac->link.speed_mask = XLGMAC_CONFIG_SS; |
1708 | |
1709 | mac->mii.addr = XGMAC_MDIO_ADDR; |
1710 | mac->mii.data = XGMAC_MDIO_DATA; |
1711 | mac->mii.addr_shift = 16; |
1712 | mac->mii.addr_mask = GENMASK(20, 16); |
1713 | mac->mii.reg_shift = 0; |
1714 | mac->mii.reg_mask = GENMASK(15, 0); |
1715 | mac->mii.clk_csr_shift = 19; |
1716 | mac->mii.clk_csr_mask = GENMASK(21, 19); |
1717 | |
1718 | return 0; |
1719 | } |
1720 | |