1  // SPDX-License-Identifier: (GPL-2.0 OR MIT)
2  /*
3   * Copyright (c) 2018 Synopsys, Inc. and/or its affiliates.
4   * stmmac XGMAC support.
5   */
6  
7  #include <linux/bitrev.h>
8  #include <linux/crc32.h>
9  #include <linux/iopoll.h>
10  #include "stmmac.h"
11  #include "stmmac_ptp.h"
12  #include "dwxlgmac2.h"
13  #include "dwxgmac2.h"
14  
dwxgmac2_core_init(struct mac_device_info * hw,struct net_device * dev)15  static void dwxgmac2_core_init(struct mac_device_info *hw,
16  			       struct net_device *dev)
17  {
18  	void __iomem *ioaddr = hw->pcsr;
19  	u32 tx, rx;
20  
21  	tx = readl(ioaddr + XGMAC_TX_CONFIG);
22  	rx = readl(ioaddr + XGMAC_RX_CONFIG);
23  
24  	tx |= XGMAC_CORE_INIT_TX;
25  	rx |= XGMAC_CORE_INIT_RX;
26  
27  	if (hw->ps) {
28  		tx |= XGMAC_CONFIG_TE;
29  		tx &= ~hw->link.speed_mask;
30  
31  		switch (hw->ps) {
32  		case SPEED_10000:
33  			tx |= hw->link.xgmii.speed10000;
34  			break;
35  		case SPEED_2500:
36  			tx |= hw->link.speed2500;
37  			break;
38  		case SPEED_1000:
39  		default:
40  			tx |= hw->link.speed1000;
41  			break;
42  		}
43  	}
44  
45  	writel(tx, ioaddr + XGMAC_TX_CONFIG);
46  	writel(rx, ioaddr + XGMAC_RX_CONFIG);
47  	writel(XGMAC_INT_DEFAULT_EN, ioaddr + XGMAC_INT_EN);
48  }
49  
dwxgmac2_set_mac(void __iomem * ioaddr,bool enable)50  static void dwxgmac2_set_mac(void __iomem *ioaddr, bool enable)
51  {
52  	u32 tx = readl(ioaddr + XGMAC_TX_CONFIG);
53  	u32 rx = readl(ioaddr + XGMAC_RX_CONFIG);
54  
55  	if (enable) {
56  		tx |= XGMAC_CONFIG_TE;
57  		rx |= XGMAC_CONFIG_RE;
58  	} else {
59  		tx &= ~XGMAC_CONFIG_TE;
60  		rx &= ~XGMAC_CONFIG_RE;
61  	}
62  
63  	writel(tx, ioaddr + XGMAC_TX_CONFIG);
64  	writel(rx, ioaddr + XGMAC_RX_CONFIG);
65  }
66  
dwxgmac2_rx_ipc(struct mac_device_info * hw)67  static int dwxgmac2_rx_ipc(struct mac_device_info *hw)
68  {
69  	void __iomem *ioaddr = hw->pcsr;
70  	u32 value;
71  
72  	value = readl(ioaddr + XGMAC_RX_CONFIG);
73  	if (hw->rx_csum)
74  		value |= XGMAC_CONFIG_IPC;
75  	else
76  		value &= ~XGMAC_CONFIG_IPC;
77  	writel(value, ioaddr + XGMAC_RX_CONFIG);
78  
79  	return !!(readl(ioaddr + XGMAC_RX_CONFIG) & XGMAC_CONFIG_IPC);
80  }
81  
dwxgmac2_rx_queue_enable(struct mac_device_info * hw,u8 mode,u32 queue)82  static void dwxgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode,
83  				     u32 queue)
84  {
85  	void __iomem *ioaddr = hw->pcsr;
86  	u32 value;
87  
88  	value = readl(ioaddr + XGMAC_RXQ_CTRL0) & ~XGMAC_RXQEN(queue);
89  	if (mode == MTL_QUEUE_AVB)
90  		value |= 0x1 << XGMAC_RXQEN_SHIFT(queue);
91  	else if (mode == MTL_QUEUE_DCB)
92  		value |= 0x2 << XGMAC_RXQEN_SHIFT(queue);
93  	writel(value, ioaddr + XGMAC_RXQ_CTRL0);
94  }
95  
dwxgmac2_rx_queue_prio(struct mac_device_info * hw,u32 prio,u32 queue)96  static void dwxgmac2_rx_queue_prio(struct mac_device_info *hw, u32 prio,
97  				   u32 queue)
98  {
99  	void __iomem *ioaddr = hw->pcsr;
100  	u32 clear_mask = 0;
101  	u32 ctrl2, ctrl3;
102  	int i;
103  
104  	ctrl2 = readl(ioaddr + XGMAC_RXQ_CTRL2);
105  	ctrl3 = readl(ioaddr + XGMAC_RXQ_CTRL3);
106  
107  	/* The software must ensure that the same priority
108  	 * is not mapped to multiple Rx queues
109  	 */
110  	for (i = 0; i < 4; i++)
111  		clear_mask |= ((prio << XGMAC_PSRQ_SHIFT(i)) &
112  						XGMAC_PSRQ(i));
113  
114  	ctrl2 &= ~clear_mask;
115  	ctrl3 &= ~clear_mask;
116  
117  	/* First assign new priorities to a queue, then
118  	 * clear them from others queues
119  	 */
120  	if (queue < 4) {
121  		ctrl2 |= (prio << XGMAC_PSRQ_SHIFT(queue)) &
122  						XGMAC_PSRQ(queue);
123  
124  		writel(ctrl2, ioaddr + XGMAC_RXQ_CTRL2);
125  		writel(ctrl3, ioaddr + XGMAC_RXQ_CTRL3);
126  	} else {
127  		queue -= 4;
128  
129  		ctrl3 |= (prio << XGMAC_PSRQ_SHIFT(queue)) &
130  						XGMAC_PSRQ(queue);
131  
132  		writel(ctrl3, ioaddr + XGMAC_RXQ_CTRL3);
133  		writel(ctrl2, ioaddr + XGMAC_RXQ_CTRL2);
134  	}
135  }
136  
dwxgmac2_tx_queue_prio(struct mac_device_info * hw,u32 prio,u32 queue)137  static void dwxgmac2_tx_queue_prio(struct mac_device_info *hw, u32 prio,
138  				   u32 queue)
139  {
140  	void __iomem *ioaddr = hw->pcsr;
141  	u32 value, reg;
142  
143  	reg = (queue < 4) ? XGMAC_TC_PRTY_MAP0 : XGMAC_TC_PRTY_MAP1;
144  	if (queue >= 4)
145  		queue -= 4;
146  
147  	value = readl(ioaddr + reg);
148  	value &= ~XGMAC_PSTC(queue);
149  	value |= (prio << XGMAC_PSTC_SHIFT(queue)) & XGMAC_PSTC(queue);
150  
151  	writel(value, ioaddr + reg);
152  }
153  
dwxgmac2_rx_queue_routing(struct mac_device_info * hw,u8 packet,u32 queue)154  static void dwxgmac2_rx_queue_routing(struct mac_device_info *hw,
155  				      u8 packet, u32 queue)
156  {
157  	void __iomem *ioaddr = hw->pcsr;
158  	u32 value;
159  
160  	static const struct stmmac_rx_routing dwxgmac2_route_possibilities[] = {
161  		{ XGMAC_AVCPQ, XGMAC_AVCPQ_SHIFT },
162  		{ XGMAC_PTPQ, XGMAC_PTPQ_SHIFT },
163  		{ XGMAC_DCBCPQ, XGMAC_DCBCPQ_SHIFT },
164  		{ XGMAC_UPQ, XGMAC_UPQ_SHIFT },
165  		{ XGMAC_MCBCQ, XGMAC_MCBCQ_SHIFT },
166  	};
167  
168  	value = readl(ioaddr + XGMAC_RXQ_CTRL1);
169  
170  	/* routing configuration */
171  	value &= ~dwxgmac2_route_possibilities[packet - 1].reg_mask;
172  	value |= (queue << dwxgmac2_route_possibilities[packet - 1].reg_shift) &
173  		 dwxgmac2_route_possibilities[packet - 1].reg_mask;
174  
175  	/* some packets require extra ops */
176  	if (packet == PACKET_AVCPQ)
177  		value |= FIELD_PREP(XGMAC_TACPQE, 1);
178  	else if (packet == PACKET_MCBCQ)
179  		value |= FIELD_PREP(XGMAC_MCBCQEN, 1);
180  
181  	writel(value, ioaddr + XGMAC_RXQ_CTRL1);
182  }
183  
dwxgmac2_prog_mtl_rx_algorithms(struct mac_device_info * hw,u32 rx_alg)184  static void dwxgmac2_prog_mtl_rx_algorithms(struct mac_device_info *hw,
185  					    u32 rx_alg)
186  {
187  	void __iomem *ioaddr = hw->pcsr;
188  	u32 value;
189  
190  	value = readl(ioaddr + XGMAC_MTL_OPMODE);
191  	value &= ~XGMAC_RAA;
192  
193  	switch (rx_alg) {
194  	case MTL_RX_ALGORITHM_SP:
195  		break;
196  	case MTL_RX_ALGORITHM_WSP:
197  		value |= XGMAC_RAA;
198  		break;
199  	default:
200  		break;
201  	}
202  
203  	writel(value, ioaddr + XGMAC_MTL_OPMODE);
204  }
205  
dwxgmac2_prog_mtl_tx_algorithms(struct mac_device_info * hw,u32 tx_alg)206  static void dwxgmac2_prog_mtl_tx_algorithms(struct mac_device_info *hw,
207  					    u32 tx_alg)
208  {
209  	void __iomem *ioaddr = hw->pcsr;
210  	bool ets = true;
211  	u32 value;
212  	int i;
213  
214  	value = readl(ioaddr + XGMAC_MTL_OPMODE);
215  	value &= ~XGMAC_ETSALG;
216  
217  	switch (tx_alg) {
218  	case MTL_TX_ALGORITHM_WRR:
219  		value |= XGMAC_WRR;
220  		break;
221  	case MTL_TX_ALGORITHM_WFQ:
222  		value |= XGMAC_WFQ;
223  		break;
224  	case MTL_TX_ALGORITHM_DWRR:
225  		value |= XGMAC_DWRR;
226  		break;
227  	default:
228  		ets = false;
229  		break;
230  	}
231  
232  	writel(value, ioaddr + XGMAC_MTL_OPMODE);
233  
234  	/* Set ETS if desired */
235  	for (i = 0; i < MTL_MAX_TX_QUEUES; i++) {
236  		value = readl(ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(i));
237  		value &= ~XGMAC_TSA;
238  		if (ets)
239  			value |= XGMAC_ETS;
240  		writel(value, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(i));
241  	}
242  }
243  
dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv * priv,struct mac_device_info * hw,u32 weight,u32 queue)244  static void dwxgmac2_set_mtl_tx_queue_weight(struct stmmac_priv *priv,
245  					     struct mac_device_info *hw,
246  					     u32 weight, u32 queue)
247  {
248  	void __iomem *ioaddr = hw->pcsr;
249  
250  	writel(weight, ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue));
251  }
252  
dwxgmac2_map_mtl_to_dma(struct mac_device_info * hw,u32 queue,u32 chan)253  static void dwxgmac2_map_mtl_to_dma(struct mac_device_info *hw, u32 queue,
254  				    u32 chan)
255  {
256  	void __iomem *ioaddr = hw->pcsr;
257  	u32 value, reg;
258  
259  	reg = (queue < 4) ? XGMAC_MTL_RXQ_DMA_MAP0 : XGMAC_MTL_RXQ_DMA_MAP1;
260  	if (queue >= 4)
261  		queue -= 4;
262  
263  	value = readl(ioaddr + reg);
264  	value &= ~XGMAC_QxMDMACH(queue);
265  	value |= (chan << XGMAC_QxMDMACH_SHIFT(queue)) & XGMAC_QxMDMACH(queue);
266  
267  	writel(value, ioaddr + reg);
268  }
269  
dwxgmac2_config_cbs(struct stmmac_priv * priv,struct mac_device_info * hw,u32 send_slope,u32 idle_slope,u32 high_credit,u32 low_credit,u32 queue)270  static void dwxgmac2_config_cbs(struct stmmac_priv *priv,
271  				struct mac_device_info *hw,
272  				u32 send_slope, u32 idle_slope,
273  				u32 high_credit, u32 low_credit, u32 queue)
274  {
275  	void __iomem *ioaddr = hw->pcsr;
276  	u32 value;
277  
278  	writel(send_slope, ioaddr + XGMAC_MTL_TCx_SENDSLOPE(queue));
279  	writel(idle_slope, ioaddr + XGMAC_MTL_TCx_QUANTUM_WEIGHT(queue));
280  	writel(high_credit, ioaddr + XGMAC_MTL_TCx_HICREDIT(queue));
281  	writel(low_credit, ioaddr + XGMAC_MTL_TCx_LOCREDIT(queue));
282  
283  	value = readl(ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue));
284  	value &= ~XGMAC_TSA;
285  	value |= XGMAC_CC | XGMAC_CBS;
286  	writel(value, ioaddr + XGMAC_MTL_TCx_ETS_CONTROL(queue));
287  }
288  
dwxgmac2_dump_regs(struct mac_device_info * hw,u32 * reg_space)289  static void dwxgmac2_dump_regs(struct mac_device_info *hw, u32 *reg_space)
290  {
291  	void __iomem *ioaddr = hw->pcsr;
292  	int i;
293  
294  	for (i = 0; i < XGMAC_MAC_REGSIZE; i++)
295  		reg_space[i] = readl(ioaddr + i * 4);
296  }
297  
dwxgmac2_host_irq_status(struct mac_device_info * hw,struct stmmac_extra_stats * x)298  static int dwxgmac2_host_irq_status(struct mac_device_info *hw,
299  				    struct stmmac_extra_stats *x)
300  {
301  	void __iomem *ioaddr = hw->pcsr;
302  	u32 stat, en;
303  	int ret = 0;
304  
305  	en = readl(ioaddr + XGMAC_INT_EN);
306  	stat = readl(ioaddr + XGMAC_INT_STATUS);
307  
308  	stat &= en;
309  
310  	if (stat & XGMAC_PMTIS) {
311  		x->irq_receive_pmt_irq_n++;
312  		readl(ioaddr + XGMAC_PMT);
313  	}
314  
315  	if (stat & XGMAC_LPIIS) {
316  		u32 lpi = readl(ioaddr + XGMAC_LPI_CTRL);
317  
318  		if (lpi & XGMAC_TLPIEN) {
319  			ret |= CORE_IRQ_TX_PATH_IN_LPI_MODE;
320  			x->irq_tx_path_in_lpi_mode_n++;
321  		}
322  		if (lpi & XGMAC_TLPIEX) {
323  			ret |= CORE_IRQ_TX_PATH_EXIT_LPI_MODE;
324  			x->irq_tx_path_exit_lpi_mode_n++;
325  		}
326  		if (lpi & XGMAC_RLPIEN)
327  			x->irq_rx_path_in_lpi_mode_n++;
328  		if (lpi & XGMAC_RLPIEX)
329  			x->irq_rx_path_exit_lpi_mode_n++;
330  	}
331  
332  	return ret;
333  }
334  
dwxgmac2_host_mtl_irq_status(struct stmmac_priv * priv,struct mac_device_info * hw,u32 chan)335  static int dwxgmac2_host_mtl_irq_status(struct stmmac_priv *priv,
336  					struct mac_device_info *hw, u32 chan)
337  {
338  	void __iomem *ioaddr = hw->pcsr;
339  	int ret = 0;
340  	u32 status;
341  
342  	status = readl(ioaddr + XGMAC_MTL_INT_STATUS);
343  	if (status & BIT(chan)) {
344  		u32 chan_status = readl(ioaddr + XGMAC_MTL_QINT_STATUS(chan));
345  
346  		if (chan_status & XGMAC_RXOVFIS)
347  			ret |= CORE_IRQ_MTL_RX_OVERFLOW;
348  
349  		writel(~0x0, ioaddr + XGMAC_MTL_QINT_STATUS(chan));
350  	}
351  
352  	return ret;
353  }
354  
dwxgmac2_flow_ctrl(struct mac_device_info * hw,unsigned int duplex,unsigned int fc,unsigned int pause_time,u32 tx_cnt)355  static void dwxgmac2_flow_ctrl(struct mac_device_info *hw, unsigned int duplex,
356  			       unsigned int fc, unsigned int pause_time,
357  			       u32 tx_cnt)
358  {
359  	void __iomem *ioaddr = hw->pcsr;
360  	u32 i;
361  
362  	if (fc & FLOW_RX)
363  		writel(XGMAC_RFE, ioaddr + XGMAC_RX_FLOW_CTRL);
364  	if (fc & FLOW_TX) {
365  		for (i = 0; i < tx_cnt; i++) {
366  			u32 value = XGMAC_TFE;
367  
368  			if (duplex)
369  				value |= pause_time << XGMAC_PT_SHIFT;
370  
371  			writel(value, ioaddr + XGMAC_Qx_TX_FLOW_CTRL(i));
372  		}
373  	}
374  }
375  
dwxgmac2_pmt(struct mac_device_info * hw,unsigned long mode)376  static void dwxgmac2_pmt(struct mac_device_info *hw, unsigned long mode)
377  {
378  	void __iomem *ioaddr = hw->pcsr;
379  	u32 val = 0x0;
380  
381  	if (mode & WAKE_MAGIC)
382  		val |= XGMAC_PWRDWN | XGMAC_MGKPKTEN;
383  	if (mode & WAKE_UCAST)
384  		val |= XGMAC_PWRDWN | XGMAC_GLBLUCAST | XGMAC_RWKPKTEN;
385  	if (val) {
386  		u32 cfg = readl(ioaddr + XGMAC_RX_CONFIG);
387  		cfg |= XGMAC_CONFIG_RE;
388  		writel(cfg, ioaddr + XGMAC_RX_CONFIG);
389  	}
390  
391  	writel(val, ioaddr + XGMAC_PMT);
392  }
393  
dwxgmac2_set_umac_addr(struct mac_device_info * hw,const unsigned char * addr,unsigned int reg_n)394  static void dwxgmac2_set_umac_addr(struct mac_device_info *hw,
395  				   const unsigned char *addr,
396  				   unsigned int reg_n)
397  {
398  	void __iomem *ioaddr = hw->pcsr;
399  	u32 value;
400  
401  	value = (addr[5] << 8) | addr[4];
402  	writel(value | XGMAC_AE, ioaddr + XGMAC_ADDRx_HIGH(reg_n));
403  
404  	value = (addr[3] << 24) | (addr[2] << 16) | (addr[1] << 8) | addr[0];
405  	writel(value, ioaddr + XGMAC_ADDRx_LOW(reg_n));
406  }
407  
dwxgmac2_get_umac_addr(struct mac_device_info * hw,unsigned char * addr,unsigned int reg_n)408  static void dwxgmac2_get_umac_addr(struct mac_device_info *hw,
409  				   unsigned char *addr, unsigned int reg_n)
410  {
411  	void __iomem *ioaddr = hw->pcsr;
412  	u32 hi_addr, lo_addr;
413  
414  	/* Read the MAC address from the hardware */
415  	hi_addr = readl(ioaddr + XGMAC_ADDRx_HIGH(reg_n));
416  	lo_addr = readl(ioaddr + XGMAC_ADDRx_LOW(reg_n));
417  
418  	/* Extract the MAC address from the high and low words */
419  	addr[0] = lo_addr & 0xff;
420  	addr[1] = (lo_addr >> 8) & 0xff;
421  	addr[2] = (lo_addr >> 16) & 0xff;
422  	addr[3] = (lo_addr >> 24) & 0xff;
423  	addr[4] = hi_addr & 0xff;
424  	addr[5] = (hi_addr >> 8) & 0xff;
425  }
426  
dwxgmac2_set_eee_mode(struct mac_device_info * hw,bool en_tx_lpi_clockgating)427  static void dwxgmac2_set_eee_mode(struct mac_device_info *hw,
428  				  bool en_tx_lpi_clockgating)
429  {
430  	void __iomem *ioaddr = hw->pcsr;
431  	u32 value;
432  
433  	value = readl(ioaddr + XGMAC_LPI_CTRL);
434  
435  	value |= XGMAC_LPITXEN | XGMAC_LPITXA;
436  	if (en_tx_lpi_clockgating)
437  		value |= XGMAC_TXCGE;
438  
439  	writel(value, ioaddr + XGMAC_LPI_CTRL);
440  }
441  
dwxgmac2_reset_eee_mode(struct mac_device_info * hw)442  static void dwxgmac2_reset_eee_mode(struct mac_device_info *hw)
443  {
444  	void __iomem *ioaddr = hw->pcsr;
445  	u32 value;
446  
447  	value = readl(ioaddr + XGMAC_LPI_CTRL);
448  	value &= ~(XGMAC_LPITXEN | XGMAC_LPITXA | XGMAC_TXCGE);
449  	writel(value, ioaddr + XGMAC_LPI_CTRL);
450  }
451  
dwxgmac2_set_eee_pls(struct mac_device_info * hw,int link)452  static void dwxgmac2_set_eee_pls(struct mac_device_info *hw, int link)
453  {
454  	void __iomem *ioaddr = hw->pcsr;
455  	u32 value;
456  
457  	value = readl(ioaddr + XGMAC_LPI_CTRL);
458  	if (link)
459  		value |= XGMAC_PLS;
460  	else
461  		value &= ~XGMAC_PLS;
462  	writel(value, ioaddr + XGMAC_LPI_CTRL);
463  }
464  
dwxgmac2_set_eee_timer(struct mac_device_info * hw,int ls,int tw)465  static void dwxgmac2_set_eee_timer(struct mac_device_info *hw, int ls, int tw)
466  {
467  	void __iomem *ioaddr = hw->pcsr;
468  	u32 value;
469  
470  	value = (tw & 0xffff) | ((ls & 0x3ff) << 16);
471  	writel(value, ioaddr + XGMAC_LPI_TIMER_CTRL);
472  }
473  
dwxgmac2_set_mchash(void __iomem * ioaddr,u32 * mcfilterbits,int mcbitslog2)474  static void dwxgmac2_set_mchash(void __iomem *ioaddr, u32 *mcfilterbits,
475  				int mcbitslog2)
476  {
477  	int numhashregs, regs;
478  
479  	switch (mcbitslog2) {
480  	case 6:
481  		numhashregs = 2;
482  		break;
483  	case 7:
484  		numhashregs = 4;
485  		break;
486  	case 8:
487  		numhashregs = 8;
488  		break;
489  	default:
490  		return;
491  	}
492  
493  	for (regs = 0; regs < numhashregs; regs++)
494  		writel(mcfilterbits[regs], ioaddr + XGMAC_HASH_TABLE(regs));
495  }
496  
dwxgmac2_set_filter(struct mac_device_info * hw,struct net_device * dev)497  static void dwxgmac2_set_filter(struct mac_device_info *hw,
498  				struct net_device *dev)
499  {
500  	void __iomem *ioaddr = (void __iomem *)dev->base_addr;
501  	u32 value = readl(ioaddr + XGMAC_PACKET_FILTER);
502  	int mcbitslog2 = hw->mcast_bits_log2;
503  	u32 mc_filter[8];
504  	int i;
505  
506  	value &= ~(XGMAC_FILTER_PR | XGMAC_FILTER_HMC | XGMAC_FILTER_PM);
507  	value |= XGMAC_FILTER_HPF;
508  
509  	memset(mc_filter, 0, sizeof(mc_filter));
510  
511  	if (dev->flags & IFF_PROMISC) {
512  		value |= XGMAC_FILTER_PR;
513  		value |= XGMAC_FILTER_PCF;
514  	} else if ((dev->flags & IFF_ALLMULTI) ||
515  		   (netdev_mc_count(dev) > hw->multicast_filter_bins)) {
516  		value |= XGMAC_FILTER_PM;
517  
518  		for (i = 0; i < XGMAC_MAX_HASH_TABLE; i++)
519  			writel(~0x0, ioaddr + XGMAC_HASH_TABLE(i));
520  	} else if (!netdev_mc_empty(dev) && (dev->flags & IFF_MULTICAST)) {
521  		struct netdev_hw_addr *ha;
522  
523  		value |= XGMAC_FILTER_HMC;
524  
525  		netdev_for_each_mc_addr(ha, dev) {
526  			u32 nr = (bitrev32(~crc32_le(~0, ha->addr, 6)) >>
527  					(32 - mcbitslog2));
528  			mc_filter[nr >> 5] |= (1 << (nr & 0x1F));
529  		}
530  	}
531  
532  	dwxgmac2_set_mchash(ioaddr, mc_filter, mcbitslog2);
533  
534  	/* Handle multiple unicast addresses */
535  	if (netdev_uc_count(dev) > hw->unicast_filter_entries) {
536  		value |= XGMAC_FILTER_PR;
537  	} else {
538  		struct netdev_hw_addr *ha;
539  		int reg = 1;
540  
541  		netdev_for_each_uc_addr(ha, dev) {
542  			dwxgmac2_set_umac_addr(hw, ha->addr, reg);
543  			reg++;
544  		}
545  
546  		for ( ; reg < XGMAC_ADDR_MAX; reg++) {
547  			writel(0, ioaddr + XGMAC_ADDRx_HIGH(reg));
548  			writel(0, ioaddr + XGMAC_ADDRx_LOW(reg));
549  		}
550  	}
551  
552  	writel(value, ioaddr + XGMAC_PACKET_FILTER);
553  }
554  
dwxgmac2_set_mac_loopback(void __iomem * ioaddr,bool enable)555  static void dwxgmac2_set_mac_loopback(void __iomem *ioaddr, bool enable)
556  {
557  	u32 value = readl(ioaddr + XGMAC_RX_CONFIG);
558  
559  	if (enable)
560  		value |= XGMAC_CONFIG_LM;
561  	else
562  		value &= ~XGMAC_CONFIG_LM;
563  
564  	writel(value, ioaddr + XGMAC_RX_CONFIG);
565  }
566  
dwxgmac2_rss_write_reg(void __iomem * ioaddr,bool is_key,int idx,u32 val)567  static int dwxgmac2_rss_write_reg(void __iomem *ioaddr, bool is_key, int idx,
568  				  u32 val)
569  {
570  	u32 ctrl = 0;
571  
572  	writel(val, ioaddr + XGMAC_RSS_DATA);
573  	ctrl |= idx << XGMAC_RSSIA_SHIFT;
574  	ctrl |= is_key ? XGMAC_ADDRT : 0x0;
575  	ctrl |= XGMAC_OB;
576  	writel(ctrl, ioaddr + XGMAC_RSS_ADDR);
577  
578  	return readl_poll_timeout(ioaddr + XGMAC_RSS_ADDR, ctrl,
579  				  !(ctrl & XGMAC_OB), 100, 10000);
580  }
581  
dwxgmac2_rss_configure(struct mac_device_info * hw,struct stmmac_rss * cfg,u32 num_rxq)582  static int dwxgmac2_rss_configure(struct mac_device_info *hw,
583  				  struct stmmac_rss *cfg, u32 num_rxq)
584  {
585  	void __iomem *ioaddr = hw->pcsr;
586  	u32 value, *key;
587  	int i, ret;
588  
589  	value = readl(ioaddr + XGMAC_RSS_CTRL);
590  	if (!cfg || !cfg->enable) {
591  		value &= ~XGMAC_RSSE;
592  		writel(value, ioaddr + XGMAC_RSS_CTRL);
593  		return 0;
594  	}
595  
596  	key = (u32 *)cfg->key;
597  	for (i = 0; i < (ARRAY_SIZE(cfg->key) / sizeof(u32)); i++) {
598  		ret = dwxgmac2_rss_write_reg(ioaddr, true, i, key[i]);
599  		if (ret)
600  			return ret;
601  	}
602  
603  	for (i = 0; i < ARRAY_SIZE(cfg->table); i++) {
604  		ret = dwxgmac2_rss_write_reg(ioaddr, false, i, cfg->table[i]);
605  		if (ret)
606  			return ret;
607  	}
608  
609  	for (i = 0; i < num_rxq; i++)
610  		dwxgmac2_map_mtl_to_dma(hw, i, XGMAC_QDDMACH);
611  
612  	value |= XGMAC_UDP4TE | XGMAC_TCP4TE | XGMAC_IP2TE | XGMAC_RSSE;
613  	writel(value, ioaddr + XGMAC_RSS_CTRL);
614  	return 0;
615  }
616  
dwxgmac2_update_vlan_hash(struct mac_device_info * hw,u32 hash,u16 perfect_match,bool is_double)617  static void dwxgmac2_update_vlan_hash(struct mac_device_info *hw, u32 hash,
618  				      u16 perfect_match, bool is_double)
619  {
620  	void __iomem *ioaddr = hw->pcsr;
621  
622  	writel(hash, ioaddr + XGMAC_VLAN_HASH_TABLE);
623  
624  	if (hash) {
625  		u32 value = readl(ioaddr + XGMAC_PACKET_FILTER);
626  
627  		value |= XGMAC_FILTER_VTFE;
628  
629  		writel(value, ioaddr + XGMAC_PACKET_FILTER);
630  
631  		value = readl(ioaddr + XGMAC_VLAN_TAG);
632  
633  		value |= XGMAC_VLAN_VTHM | XGMAC_VLAN_ETV;
634  		if (is_double) {
635  			value |= XGMAC_VLAN_EDVLP;
636  			value |= XGMAC_VLAN_ESVL;
637  			value |= XGMAC_VLAN_DOVLTC;
638  		} else {
639  			value &= ~XGMAC_VLAN_EDVLP;
640  			value &= ~XGMAC_VLAN_ESVL;
641  			value &= ~XGMAC_VLAN_DOVLTC;
642  		}
643  
644  		value &= ~XGMAC_VLAN_VID;
645  		writel(value, ioaddr + XGMAC_VLAN_TAG);
646  	} else if (perfect_match) {
647  		u32 value = readl(ioaddr + XGMAC_PACKET_FILTER);
648  
649  		value |= XGMAC_FILTER_VTFE;
650  
651  		writel(value, ioaddr + XGMAC_PACKET_FILTER);
652  
653  		value = readl(ioaddr + XGMAC_VLAN_TAG);
654  
655  		value &= ~XGMAC_VLAN_VTHM;
656  		value |= XGMAC_VLAN_ETV;
657  		if (is_double) {
658  			value |= XGMAC_VLAN_EDVLP;
659  			value |= XGMAC_VLAN_ESVL;
660  			value |= XGMAC_VLAN_DOVLTC;
661  		} else {
662  			value &= ~XGMAC_VLAN_EDVLP;
663  			value &= ~XGMAC_VLAN_ESVL;
664  			value &= ~XGMAC_VLAN_DOVLTC;
665  		}
666  
667  		value &= ~XGMAC_VLAN_VID;
668  		writel(value | perfect_match, ioaddr + XGMAC_VLAN_TAG);
669  	} else {
670  		u32 value = readl(ioaddr + XGMAC_PACKET_FILTER);
671  
672  		value &= ~XGMAC_FILTER_VTFE;
673  
674  		writel(value, ioaddr + XGMAC_PACKET_FILTER);
675  
676  		value = readl(ioaddr + XGMAC_VLAN_TAG);
677  
678  		value &= ~(XGMAC_VLAN_VTHM | XGMAC_VLAN_ETV);
679  		value &= ~(XGMAC_VLAN_EDVLP | XGMAC_VLAN_ESVL);
680  		value &= ~XGMAC_VLAN_DOVLTC;
681  		value &= ~XGMAC_VLAN_VID;
682  
683  		writel(value, ioaddr + XGMAC_VLAN_TAG);
684  	}
685  }
686  
687  struct dwxgmac3_error_desc {
688  	bool valid;
689  	const char *desc;
690  	const char *detailed_desc;
691  };
692  
693  #define STAT_OFF(field)		offsetof(struct stmmac_safety_stats, field)
694  
dwxgmac3_log_error(struct net_device * ndev,u32 value,bool corr,const char * module_name,const struct dwxgmac3_error_desc * desc,unsigned long field_offset,struct stmmac_safety_stats * stats)695  static void dwxgmac3_log_error(struct net_device *ndev, u32 value, bool corr,
696  			       const char *module_name,
697  			       const struct dwxgmac3_error_desc *desc,
698  			       unsigned long field_offset,
699  			       struct stmmac_safety_stats *stats)
700  {
701  	unsigned long loc, mask;
702  	u8 *bptr = (u8 *)stats;
703  	unsigned long *ptr;
704  
705  	ptr = (unsigned long *)(bptr + field_offset);
706  
707  	mask = value;
708  	for_each_set_bit(loc, &mask, 32) {
709  		netdev_err(ndev, "Found %s error in %s: '%s: %s'\n", corr ?
710  				"correctable" : "uncorrectable", module_name,
711  				desc[loc].desc, desc[loc].detailed_desc);
712  
713  		/* Update counters */
714  		ptr[loc]++;
715  	}
716  }
717  
718  static const struct dwxgmac3_error_desc dwxgmac3_mac_errors[32]= {
719  	{ true, "ATPES", "Application Transmit Interface Parity Check Error" },
720  	{ true, "DPES", "Descriptor Cache Data Path Parity Check Error" },
721  	{ true, "TPES", "TSO Data Path Parity Check Error" },
722  	{ true, "TSOPES", "TSO Header Data Path Parity Check Error" },
723  	{ true, "MTPES", "MTL Data Path Parity Check Error" },
724  	{ true, "MTSPES", "MTL TX Status Data Path Parity Check Error" },
725  	{ true, "MTBUPES", "MAC TBU Data Path Parity Check Error" },
726  	{ true, "MTFCPES", "MAC TFC Data Path Parity Check Error" },
727  	{ true, "ARPES", "Application Receive Interface Data Path Parity Check Error" },
728  	{ true, "MRWCPES", "MTL RWC Data Path Parity Check Error" },
729  	{ true, "MRRCPES", "MTL RCC Data Path Parity Check Error" },
730  	{ true, "CWPES", "CSR Write Data Path Parity Check Error" },
731  	{ true, "ASRPES", "AXI Slave Read Data Path Parity Check Error" },
732  	{ true, "TTES", "TX FSM Timeout Error" },
733  	{ true, "RTES", "RX FSM Timeout Error" },
734  	{ true, "CTES", "CSR FSM Timeout Error" },
735  	{ true, "ATES", "APP FSM Timeout Error" },
736  	{ true, "PTES", "PTP FSM Timeout Error" },
737  	{ false, "UNKNOWN", "Unknown Error" }, /* 18 */
738  	{ false, "UNKNOWN", "Unknown Error" }, /* 19 */
739  	{ false, "UNKNOWN", "Unknown Error" }, /* 20 */
740  	{ true, "MSTTES", "Master Read/Write Timeout Error" },
741  	{ true, "SLVTES", "Slave Read/Write Timeout Error" },
742  	{ true, "ATITES", "Application Timeout on ATI Interface Error" },
743  	{ true, "ARITES", "Application Timeout on ARI Interface Error" },
744  	{ true, "FSMPES", "FSM State Parity Error" },
745  	{ false, "UNKNOWN", "Unknown Error" }, /* 26 */
746  	{ false, "UNKNOWN", "Unknown Error" }, /* 27 */
747  	{ false, "UNKNOWN", "Unknown Error" }, /* 28 */
748  	{ false, "UNKNOWN", "Unknown Error" }, /* 29 */
749  	{ false, "UNKNOWN", "Unknown Error" }, /* 30 */
750  	{ true, "CPI", "Control Register Parity Check Error" },
751  };
752  
dwxgmac3_handle_mac_err(struct net_device * ndev,void __iomem * ioaddr,bool correctable,struct stmmac_safety_stats * stats)753  static void dwxgmac3_handle_mac_err(struct net_device *ndev,
754  				    void __iomem *ioaddr, bool correctable,
755  				    struct stmmac_safety_stats *stats)
756  {
757  	u32 value;
758  
759  	value = readl(ioaddr + XGMAC_MAC_DPP_FSM_INT_STATUS);
760  	writel(value, ioaddr + XGMAC_MAC_DPP_FSM_INT_STATUS);
761  
762  	dwxgmac3_log_error(ndev, value, correctable, "MAC",
763  			   dwxgmac3_mac_errors, STAT_OFF(mac_errors), stats);
764  }
765  
766  static const struct dwxgmac3_error_desc dwxgmac3_mtl_errors[32]= {
767  	{ true, "TXCES", "MTL TX Memory Error" },
768  	{ true, "TXAMS", "MTL TX Memory Address Mismatch Error" },
769  	{ true, "TXUES", "MTL TX Memory Error" },
770  	{ false, "UNKNOWN", "Unknown Error" }, /* 3 */
771  	{ true, "RXCES", "MTL RX Memory Error" },
772  	{ true, "RXAMS", "MTL RX Memory Address Mismatch Error" },
773  	{ true, "RXUES", "MTL RX Memory Error" },
774  	{ false, "UNKNOWN", "Unknown Error" }, /* 7 */
775  	{ true, "ECES", "MTL EST Memory Error" },
776  	{ true, "EAMS", "MTL EST Memory Address Mismatch Error" },
777  	{ true, "EUES", "MTL EST Memory Error" },
778  	{ false, "UNKNOWN", "Unknown Error" }, /* 11 */
779  	{ true, "RPCES", "MTL RX Parser Memory Error" },
780  	{ true, "RPAMS", "MTL RX Parser Memory Address Mismatch Error" },
781  	{ true, "RPUES", "MTL RX Parser Memory Error" },
782  	{ false, "UNKNOWN", "Unknown Error" }, /* 15 */
783  	{ false, "UNKNOWN", "Unknown Error" }, /* 16 */
784  	{ false, "UNKNOWN", "Unknown Error" }, /* 17 */
785  	{ false, "UNKNOWN", "Unknown Error" }, /* 18 */
786  	{ false, "UNKNOWN", "Unknown Error" }, /* 19 */
787  	{ false, "UNKNOWN", "Unknown Error" }, /* 20 */
788  	{ false, "UNKNOWN", "Unknown Error" }, /* 21 */
789  	{ false, "UNKNOWN", "Unknown Error" }, /* 22 */
790  	{ false, "UNKNOWN", "Unknown Error" }, /* 23 */
791  	{ false, "UNKNOWN", "Unknown Error" }, /* 24 */
792  	{ false, "UNKNOWN", "Unknown Error" }, /* 25 */
793  	{ false, "UNKNOWN", "Unknown Error" }, /* 26 */
794  	{ false, "UNKNOWN", "Unknown Error" }, /* 27 */
795  	{ false, "UNKNOWN", "Unknown Error" }, /* 28 */
796  	{ false, "UNKNOWN", "Unknown Error" }, /* 29 */
797  	{ false, "UNKNOWN", "Unknown Error" }, /* 30 */
798  	{ false, "UNKNOWN", "Unknown Error" }, /* 31 */
799  };
800  
dwxgmac3_handle_mtl_err(struct net_device * ndev,void __iomem * ioaddr,bool correctable,struct stmmac_safety_stats * stats)801  static void dwxgmac3_handle_mtl_err(struct net_device *ndev,
802  				    void __iomem *ioaddr, bool correctable,
803  				    struct stmmac_safety_stats *stats)
804  {
805  	u32 value;
806  
807  	value = readl(ioaddr + XGMAC_MTL_ECC_INT_STATUS);
808  	writel(value, ioaddr + XGMAC_MTL_ECC_INT_STATUS);
809  
810  	dwxgmac3_log_error(ndev, value, correctable, "MTL",
811  			   dwxgmac3_mtl_errors, STAT_OFF(mtl_errors), stats);
812  }
813  
814  static const struct dwxgmac3_error_desc dwxgmac3_dma_errors[32]= {
815  	{ true, "TCES", "DMA TSO Memory Error" },
816  	{ true, "TAMS", "DMA TSO Memory Address Mismatch Error" },
817  	{ true, "TUES", "DMA TSO Memory Error" },
818  	{ false, "UNKNOWN", "Unknown Error" }, /* 3 */
819  	{ true, "DCES", "DMA DCACHE Memory Error" },
820  	{ true, "DAMS", "DMA DCACHE Address Mismatch Error" },
821  	{ true, "DUES", "DMA DCACHE Memory Error" },
822  	{ false, "UNKNOWN", "Unknown Error" }, /* 7 */
823  	{ false, "UNKNOWN", "Unknown Error" }, /* 8 */
824  	{ false, "UNKNOWN", "Unknown Error" }, /* 9 */
825  	{ false, "UNKNOWN", "Unknown Error" }, /* 10 */
826  	{ false, "UNKNOWN", "Unknown Error" }, /* 11 */
827  	{ false, "UNKNOWN", "Unknown Error" }, /* 12 */
828  	{ false, "UNKNOWN", "Unknown Error" }, /* 13 */
829  	{ false, "UNKNOWN", "Unknown Error" }, /* 14 */
830  	{ false, "UNKNOWN", "Unknown Error" }, /* 15 */
831  	{ false, "UNKNOWN", "Unknown Error" }, /* 16 */
832  	{ false, "UNKNOWN", "Unknown Error" }, /* 17 */
833  	{ false, "UNKNOWN", "Unknown Error" }, /* 18 */
834  	{ false, "UNKNOWN", "Unknown Error" }, /* 19 */
835  	{ false, "UNKNOWN", "Unknown Error" }, /* 20 */
836  	{ false, "UNKNOWN", "Unknown Error" }, /* 21 */
837  	{ false, "UNKNOWN", "Unknown Error" }, /* 22 */
838  	{ false, "UNKNOWN", "Unknown Error" }, /* 23 */
839  	{ false, "UNKNOWN", "Unknown Error" }, /* 24 */
840  	{ false, "UNKNOWN", "Unknown Error" }, /* 25 */
841  	{ false, "UNKNOWN", "Unknown Error" }, /* 26 */
842  	{ false, "UNKNOWN", "Unknown Error" }, /* 27 */
843  	{ false, "UNKNOWN", "Unknown Error" }, /* 28 */
844  	{ false, "UNKNOWN", "Unknown Error" }, /* 29 */
845  	{ false, "UNKNOWN", "Unknown Error" }, /* 30 */
846  	{ false, "UNKNOWN", "Unknown Error" }, /* 31 */
847  };
848  
849  static const char dpp_rx_err[] = "Read Rx Descriptor Parity checker Error";
850  static const char dpp_tx_err[] = "Read Tx Descriptor Parity checker Error";
851  static const struct dwxgmac3_error_desc dwxgmac3_dma_dpp_errors[32] = {
852  	{ true, "TDPES0", dpp_tx_err },
853  	{ true, "TDPES1", dpp_tx_err },
854  	{ true, "TDPES2", dpp_tx_err },
855  	{ true, "TDPES3", dpp_tx_err },
856  	{ true, "TDPES4", dpp_tx_err },
857  	{ true, "TDPES5", dpp_tx_err },
858  	{ true, "TDPES6", dpp_tx_err },
859  	{ true, "TDPES7", dpp_tx_err },
860  	{ true, "TDPES8", dpp_tx_err },
861  	{ true, "TDPES9", dpp_tx_err },
862  	{ true, "TDPES10", dpp_tx_err },
863  	{ true, "TDPES11", dpp_tx_err },
864  	{ true, "TDPES12", dpp_tx_err },
865  	{ true, "TDPES13", dpp_tx_err },
866  	{ true, "TDPES14", dpp_tx_err },
867  	{ true, "TDPES15", dpp_tx_err },
868  	{ true, "RDPES0", dpp_rx_err },
869  	{ true, "RDPES1", dpp_rx_err },
870  	{ true, "RDPES2", dpp_rx_err },
871  	{ true, "RDPES3", dpp_rx_err },
872  	{ true, "RDPES4", dpp_rx_err },
873  	{ true, "RDPES5", dpp_rx_err },
874  	{ true, "RDPES6", dpp_rx_err },
875  	{ true, "RDPES7", dpp_rx_err },
876  	{ true, "RDPES8", dpp_rx_err },
877  	{ true, "RDPES9", dpp_rx_err },
878  	{ true, "RDPES10", dpp_rx_err },
879  	{ true, "RDPES11", dpp_rx_err },
880  	{ true, "RDPES12", dpp_rx_err },
881  	{ true, "RDPES13", dpp_rx_err },
882  	{ true, "RDPES14", dpp_rx_err },
883  	{ true, "RDPES15", dpp_rx_err },
884  };
885  
dwxgmac3_handle_dma_err(struct net_device * ndev,void __iomem * ioaddr,bool correctable,struct stmmac_safety_stats * stats)886  static void dwxgmac3_handle_dma_err(struct net_device *ndev,
887  				    void __iomem *ioaddr, bool correctable,
888  				    struct stmmac_safety_stats *stats)
889  {
890  	u32 value;
891  
892  	value = readl(ioaddr + XGMAC_DMA_ECC_INT_STATUS);
893  	writel(value, ioaddr + XGMAC_DMA_ECC_INT_STATUS);
894  
895  	dwxgmac3_log_error(ndev, value, correctable, "DMA",
896  			   dwxgmac3_dma_errors, STAT_OFF(dma_errors), stats);
897  
898  	value = readl(ioaddr + XGMAC_DMA_DPP_INT_STATUS);
899  	writel(value, ioaddr + XGMAC_DMA_DPP_INT_STATUS);
900  
901  	dwxgmac3_log_error(ndev, value, false, "DMA_DPP",
902  			   dwxgmac3_dma_dpp_errors,
903  			   STAT_OFF(dma_dpp_errors), stats);
904  }
905  
906  static int
dwxgmac3_safety_feat_config(void __iomem * ioaddr,unsigned int asp,struct stmmac_safety_feature_cfg * safety_cfg)907  dwxgmac3_safety_feat_config(void __iomem *ioaddr, unsigned int asp,
908  			    struct stmmac_safety_feature_cfg *safety_cfg)
909  {
910  	u32 value;
911  
912  	if (!asp)
913  		return -EINVAL;
914  
915  	/* 1. Enable Safety Features */
916  	writel(0x0, ioaddr + XGMAC_MTL_ECC_CONTROL);
917  
918  	/* 2. Enable MTL Safety Interrupts */
919  	value = readl(ioaddr + XGMAC_MTL_ECC_INT_ENABLE);
920  	value |= XGMAC_RPCEIE; /* RX Parser Memory Correctable Error */
921  	value |= XGMAC_ECEIE; /* EST Memory Correctable Error */
922  	value |= XGMAC_RXCEIE; /* RX Memory Correctable Error */
923  	value |= XGMAC_TXCEIE; /* TX Memory Correctable Error */
924  	writel(value, ioaddr + XGMAC_MTL_ECC_INT_ENABLE);
925  
926  	/* 3. Enable DMA Safety Interrupts */
927  	value = readl(ioaddr + XGMAC_DMA_ECC_INT_ENABLE);
928  	value |= XGMAC_DCEIE; /* Descriptor Cache Memory Correctable Error */
929  	value |= XGMAC_TCEIE; /* TSO Memory Correctable Error */
930  	writel(value, ioaddr + XGMAC_DMA_ECC_INT_ENABLE);
931  
932  	/* 0x2: Without ECC or Parity Ports on External Application Interface
933  	 * 0x4: Only ECC Protection for External Memory feature is selected
934  	 */
935  	if (asp == 0x2 || asp == 0x4)
936  		return 0;
937  
938  	/* 4. Enable Parity and Timeout for FSM */
939  	value = readl(ioaddr + XGMAC_MAC_FSM_CONTROL);
940  	value |= XGMAC_PRTYEN; /* FSM Parity Feature */
941  	value |= XGMAC_TMOUTEN; /* FSM Timeout Feature */
942  	writel(value, ioaddr + XGMAC_MAC_FSM_CONTROL);
943  
944  	/* 5. Enable Data Path Parity Protection */
945  	value = readl(ioaddr + XGMAC_MTL_DPP_CONTROL);
946  	/* already enabled by default, explicit enable it again */
947  	value &= ~XGMAC_DPP_DISABLE;
948  	writel(value, ioaddr + XGMAC_MTL_DPP_CONTROL);
949  
950  	return 0;
951  }
952  
dwxgmac3_safety_feat_irq_status(struct net_device * ndev,void __iomem * ioaddr,unsigned int asp,struct stmmac_safety_stats * stats)953  static int dwxgmac3_safety_feat_irq_status(struct net_device *ndev,
954  					   void __iomem *ioaddr,
955  					   unsigned int asp,
956  					   struct stmmac_safety_stats *stats)
957  {
958  	bool err, corr;
959  	u32 mtl, dma;
960  	int ret = 0;
961  
962  	if (!asp)
963  		return -EINVAL;
964  
965  	mtl = readl(ioaddr + XGMAC_MTL_SAFETY_INT_STATUS);
966  	dma = readl(ioaddr + XGMAC_DMA_SAFETY_INT_STATUS);
967  
968  	err = (mtl & XGMAC_MCSIS) || (dma & XGMAC_MCSIS);
969  	corr = false;
970  	if (err) {
971  		dwxgmac3_handle_mac_err(ndev, ioaddr, corr, stats);
972  		ret |= !corr;
973  	}
974  
975  	err = (mtl & (XGMAC_MEUIS | XGMAC_MECIS)) ||
976  	      (dma & (XGMAC_MSUIS | XGMAC_MSCIS));
977  	corr = (mtl & XGMAC_MECIS) || (dma & XGMAC_MSCIS);
978  	if (err) {
979  		dwxgmac3_handle_mtl_err(ndev, ioaddr, corr, stats);
980  		ret |= !corr;
981  	}
982  
983  	/* DMA_DPP_Interrupt_Status is indicated by MCSIS bit in
984  	 * DMA_Safety_Interrupt_Status, so we handle DMA Data Path
985  	 * Parity Errors here
986  	 */
987  	err = dma & (XGMAC_DEUIS | XGMAC_DECIS | XGMAC_MCSIS);
988  	corr = dma & XGMAC_DECIS;
989  	if (err) {
990  		dwxgmac3_handle_dma_err(ndev, ioaddr, corr, stats);
991  		ret |= !corr;
992  	}
993  
994  	return ret;
995  }
996  
997  static const struct dwxgmac3_error {
998  	const struct dwxgmac3_error_desc *desc;
999  } dwxgmac3_all_errors[] = {
1000  	{ dwxgmac3_mac_errors },
1001  	{ dwxgmac3_mtl_errors },
1002  	{ dwxgmac3_dma_errors },
1003  	{ dwxgmac3_dma_dpp_errors },
1004  };
1005  
dwxgmac3_safety_feat_dump(struct stmmac_safety_stats * stats,int index,unsigned long * count,const char ** desc)1006  static int dwxgmac3_safety_feat_dump(struct stmmac_safety_stats *stats,
1007  				     int index, unsigned long *count,
1008  				     const char **desc)
1009  {
1010  	int module = index / 32, offset = index % 32;
1011  	unsigned long *ptr = (unsigned long *)stats;
1012  
1013  	if (module >= ARRAY_SIZE(dwxgmac3_all_errors))
1014  		return -EINVAL;
1015  	if (!dwxgmac3_all_errors[module].desc[offset].valid)
1016  		return -EINVAL;
1017  	if (count)
1018  		*count = *(ptr + index);
1019  	if (desc)
1020  		*desc = dwxgmac3_all_errors[module].desc[offset].desc;
1021  	return 0;
1022  }
1023  
dwxgmac3_rxp_disable(void __iomem * ioaddr)1024  static int dwxgmac3_rxp_disable(void __iomem *ioaddr)
1025  {
1026  	u32 val = readl(ioaddr + XGMAC_MTL_OPMODE);
1027  
1028  	val &= ~XGMAC_FRPE;
1029  	writel(val, ioaddr + XGMAC_MTL_OPMODE);
1030  
1031  	return 0;
1032  }
1033  
dwxgmac3_rxp_enable(void __iomem * ioaddr)1034  static void dwxgmac3_rxp_enable(void __iomem *ioaddr)
1035  {
1036  	u32 val;
1037  
1038  	val = readl(ioaddr + XGMAC_MTL_OPMODE);
1039  	val |= XGMAC_FRPE;
1040  	writel(val, ioaddr + XGMAC_MTL_OPMODE);
1041  }
1042  
dwxgmac3_rxp_update_single_entry(void __iomem * ioaddr,struct stmmac_tc_entry * entry,int pos)1043  static int dwxgmac3_rxp_update_single_entry(void __iomem *ioaddr,
1044  					    struct stmmac_tc_entry *entry,
1045  					    int pos)
1046  {
1047  	int ret, i;
1048  
1049  	for (i = 0; i < (sizeof(entry->val) / sizeof(u32)); i++) {
1050  		int real_pos = pos * (sizeof(entry->val) / sizeof(u32)) + i;
1051  		u32 val;
1052  
1053  		/* Wait for ready */
1054  		ret = readl_poll_timeout(ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST,
1055  					 val, !(val & XGMAC_STARTBUSY), 1, 10000);
1056  		if (ret)
1057  			return ret;
1058  
1059  		/* Write data */
1060  		val = *((u32 *)&entry->val + i);
1061  		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_DATA);
1062  
1063  		/* Write pos */
1064  		val = real_pos & XGMAC_ADDR;
1065  		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST);
1066  
1067  		/* Write OP */
1068  		val |= XGMAC_WRRDN;
1069  		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST);
1070  
1071  		/* Start Write */
1072  		val |= XGMAC_STARTBUSY;
1073  		writel(val, ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST);
1074  
1075  		/* Wait for done */
1076  		ret = readl_poll_timeout(ioaddr + XGMAC_MTL_RXP_IACC_CTRL_ST,
1077  					 val, !(val & XGMAC_STARTBUSY), 1, 10000);
1078  		if (ret)
1079  			return ret;
1080  	}
1081  
1082  	return 0;
1083  }
1084  
1085  static struct stmmac_tc_entry *
dwxgmac3_rxp_get_next_entry(struct stmmac_tc_entry * entries,unsigned int count,u32 curr_prio)1086  dwxgmac3_rxp_get_next_entry(struct stmmac_tc_entry *entries,
1087  			    unsigned int count, u32 curr_prio)
1088  {
1089  	struct stmmac_tc_entry *entry;
1090  	u32 min_prio = ~0x0;
1091  	int i, min_prio_idx;
1092  	bool found = false;
1093  
1094  	for (i = count - 1; i >= 0; i--) {
1095  		entry = &entries[i];
1096  
1097  		/* Do not update unused entries */
1098  		if (!entry->in_use)
1099  			continue;
1100  		/* Do not update already updated entries (i.e. fragments) */
1101  		if (entry->in_hw)
1102  			continue;
1103  		/* Let last entry be updated last */
1104  		if (entry->is_last)
1105  			continue;
1106  		/* Do not return fragments */
1107  		if (entry->is_frag)
1108  			continue;
1109  		/* Check if we already checked this prio */
1110  		if (entry->prio < curr_prio)
1111  			continue;
1112  		/* Check if this is the minimum prio */
1113  		if (entry->prio < min_prio) {
1114  			min_prio = entry->prio;
1115  			min_prio_idx = i;
1116  			found = true;
1117  		}
1118  	}
1119  
1120  	if (found)
1121  		return &entries[min_prio_idx];
1122  	return NULL;
1123  }
1124  
dwxgmac3_rxp_config(void __iomem * ioaddr,struct stmmac_tc_entry * entries,unsigned int count)1125  static int dwxgmac3_rxp_config(void __iomem *ioaddr,
1126  			       struct stmmac_tc_entry *entries,
1127  			       unsigned int count)
1128  {
1129  	struct stmmac_tc_entry *entry, *frag;
1130  	int i, ret, nve = 0;
1131  	u32 curr_prio = 0;
1132  	u32 old_val, val;
1133  
1134  	/* Force disable RX */
1135  	old_val = readl(ioaddr + XGMAC_RX_CONFIG);
1136  	val = old_val & ~XGMAC_CONFIG_RE;
1137  	writel(val, ioaddr + XGMAC_RX_CONFIG);
1138  
1139  	/* Disable RX Parser */
1140  	ret = dwxgmac3_rxp_disable(ioaddr);
1141  	if (ret)
1142  		goto re_enable;
1143  
1144  	/* Set all entries as NOT in HW */
1145  	for (i = 0; i < count; i++) {
1146  		entry = &entries[i];
1147  		entry->in_hw = false;
1148  	}
1149  
1150  	/* Update entries by reverse order */
1151  	while (1) {
1152  		entry = dwxgmac3_rxp_get_next_entry(entries, count, curr_prio);
1153  		if (!entry)
1154  			break;
1155  
1156  		curr_prio = entry->prio;
1157  		frag = entry->frag_ptr;
1158  
1159  		/* Set special fragment requirements */
1160  		if (frag) {
1161  			entry->val.af = 0;
1162  			entry->val.rf = 0;
1163  			entry->val.nc = 1;
1164  			entry->val.ok_index = nve + 2;
1165  		}
1166  
1167  		ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve);
1168  		if (ret)
1169  			goto re_enable;
1170  
1171  		entry->table_pos = nve++;
1172  		entry->in_hw = true;
1173  
1174  		if (frag && !frag->in_hw) {
1175  			ret = dwxgmac3_rxp_update_single_entry(ioaddr, frag, nve);
1176  			if (ret)
1177  				goto re_enable;
1178  			frag->table_pos = nve++;
1179  			frag->in_hw = true;
1180  		}
1181  	}
1182  
1183  	if (!nve)
1184  		goto re_enable;
1185  
1186  	/* Update all pass entry */
1187  	for (i = 0; i < count; i++) {
1188  		entry = &entries[i];
1189  		if (!entry->is_last)
1190  			continue;
1191  
1192  		ret = dwxgmac3_rxp_update_single_entry(ioaddr, entry, nve);
1193  		if (ret)
1194  			goto re_enable;
1195  
1196  		entry->table_pos = nve++;
1197  	}
1198  
1199  	/* Assume n. of parsable entries == n. of valid entries */
1200  	val = (nve << 16) & XGMAC_NPE;
1201  	val |= nve & XGMAC_NVE;
1202  	writel(val, ioaddr + XGMAC_MTL_RXP_CONTROL_STATUS);
1203  
1204  	/* Enable RX Parser */
1205  	dwxgmac3_rxp_enable(ioaddr);
1206  
1207  re_enable:
1208  	/* Re-enable RX */
1209  	writel(old_val, ioaddr + XGMAC_RX_CONFIG);
1210  	return ret;
1211  }
1212  
dwxgmac2_get_mac_tx_timestamp(struct mac_device_info * hw,u64 * ts)1213  static int dwxgmac2_get_mac_tx_timestamp(struct mac_device_info *hw, u64 *ts)
1214  {
1215  	void __iomem *ioaddr = hw->pcsr;
1216  	u32 value;
1217  
1218  	if (readl_poll_timeout_atomic(ioaddr + XGMAC_TIMESTAMP_STATUS,
1219  				      value, value & XGMAC_TXTSC, 100, 10000))
1220  		return -EBUSY;
1221  
1222  	*ts = readl(ioaddr + XGMAC_TXTIMESTAMP_NSEC) & XGMAC_TXTSSTSLO;
1223  	*ts += readl(ioaddr + XGMAC_TXTIMESTAMP_SEC) * 1000000000ULL;
1224  	return 0;
1225  }
1226  
dwxgmac2_flex_pps_config(void __iomem * ioaddr,int index,struct stmmac_pps_cfg * cfg,bool enable,u32 sub_second_inc,u32 systime_flags)1227  static int dwxgmac2_flex_pps_config(void __iomem *ioaddr, int index,
1228  				    struct stmmac_pps_cfg *cfg, bool enable,
1229  				    u32 sub_second_inc, u32 systime_flags)
1230  {
1231  	u32 tnsec = readl(ioaddr + XGMAC_PPSx_TARGET_TIME_NSEC(index));
1232  	u32 val = readl(ioaddr + XGMAC_PPS_CONTROL);
1233  	u64 period;
1234  
1235  	if (!cfg->available)
1236  		return -EINVAL;
1237  	if (tnsec & XGMAC_TRGTBUSY0)
1238  		return -EBUSY;
1239  	if (!sub_second_inc || !systime_flags)
1240  		return -EINVAL;
1241  
1242  	val &= ~XGMAC_PPSx_MASK(index);
1243  
1244  	if (!enable) {
1245  		val |= XGMAC_PPSCMDx(index, XGMAC_PPSCMD_STOP);
1246  		writel(val, ioaddr + XGMAC_PPS_CONTROL);
1247  		return 0;
1248  	}
1249  
1250  	val |= XGMAC_PPSCMDx(index, XGMAC_PPSCMD_START);
1251  	val |= XGMAC_TRGTMODSELx(index, XGMAC_PPSCMD_START);
1252  
1253  	/* XGMAC Core has 4 PPS outputs at most.
1254  	 *
1255  	 * Prior XGMAC Core 3.20, Fixed mode or Flexible mode are selectable for
1256  	 * PPS0 only via PPSEN0. PPS{1,2,3} are in Flexible mode by default,
1257  	 * and can not be switched to Fixed mode, since PPSEN{1,2,3} are
1258  	 * read-only reserved to 0.
1259  	 * But we always set PPSEN{1,2,3} do not make things worse ;-)
1260  	 *
1261  	 * From XGMAC Core 3.20 and later, PPSEN{0,1,2,3} are writable and must
1262  	 * be set, or the PPS outputs stay in Fixed PPS mode by default.
1263  	 */
1264  	val |= XGMAC_PPSENx(index);
1265  
1266  	writel(cfg->start.tv_sec, ioaddr + XGMAC_PPSx_TARGET_TIME_SEC(index));
1267  
1268  	if (!(systime_flags & PTP_TCR_TSCTRLSSR))
1269  		cfg->start.tv_nsec = (cfg->start.tv_nsec * 1000) / 465;
1270  	writel(cfg->start.tv_nsec, ioaddr + XGMAC_PPSx_TARGET_TIME_NSEC(index));
1271  
1272  	period = cfg->period.tv_sec * 1000000000;
1273  	period += cfg->period.tv_nsec;
1274  
1275  	do_div(period, sub_second_inc);
1276  
1277  	if (period <= 1)
1278  		return -EINVAL;
1279  
1280  	writel(period - 1, ioaddr + XGMAC_PPSx_INTERVAL(index));
1281  
1282  	period >>= 1;
1283  	if (period <= 1)
1284  		return -EINVAL;
1285  
1286  	writel(period - 1, ioaddr + XGMAC_PPSx_WIDTH(index));
1287  
1288  	/* Finally, activate it */
1289  	writel(val, ioaddr + XGMAC_PPS_CONTROL);
1290  	return 0;
1291  }
1292  
dwxgmac2_sarc_configure(void __iomem * ioaddr,int val)1293  static void dwxgmac2_sarc_configure(void __iomem *ioaddr, int val)
1294  {
1295  	u32 value = readl(ioaddr + XGMAC_TX_CONFIG);
1296  
1297  	value &= ~XGMAC_CONFIG_SARC;
1298  	value |= val << XGMAC_CONFIG_SARC_SHIFT;
1299  
1300  	writel(value, ioaddr + XGMAC_TX_CONFIG);
1301  }
1302  
dwxgmac2_enable_vlan(struct mac_device_info * hw,u32 type)1303  static void dwxgmac2_enable_vlan(struct mac_device_info *hw, u32 type)
1304  {
1305  	void __iomem *ioaddr = hw->pcsr;
1306  	u32 value;
1307  
1308  	value = readl(ioaddr + XGMAC_VLAN_INCL);
1309  	value |= XGMAC_VLAN_VLTI;
1310  	value |= XGMAC_VLAN_CSVL; /* Only use SVLAN */
1311  	value &= ~XGMAC_VLAN_VLC;
1312  	value |= (type << XGMAC_VLAN_VLC_SHIFT) & XGMAC_VLAN_VLC;
1313  	writel(value, ioaddr + XGMAC_VLAN_INCL);
1314  }
1315  
dwxgmac2_filter_wait(struct mac_device_info * hw)1316  static int dwxgmac2_filter_wait(struct mac_device_info *hw)
1317  {
1318  	void __iomem *ioaddr = hw->pcsr;
1319  	u32 value;
1320  
1321  	if (readl_poll_timeout(ioaddr + XGMAC_L3L4_ADDR_CTRL, value,
1322  			       !(value & XGMAC_XB), 100, 10000))
1323  		return -EBUSY;
1324  	return 0;
1325  }
1326  
dwxgmac2_filter_read(struct mac_device_info * hw,u32 filter_no,u8 reg,u32 * data)1327  static int dwxgmac2_filter_read(struct mac_device_info *hw, u32 filter_no,
1328  				u8 reg, u32 *data)
1329  {
1330  	void __iomem *ioaddr = hw->pcsr;
1331  	u32 value;
1332  	int ret;
1333  
1334  	ret = dwxgmac2_filter_wait(hw);
1335  	if (ret)
1336  		return ret;
1337  
1338  	value = ((filter_no << XGMAC_IDDR_FNUM) | reg) << XGMAC_IDDR_SHIFT;
1339  	value |= XGMAC_TT | XGMAC_XB;
1340  	writel(value, ioaddr + XGMAC_L3L4_ADDR_CTRL);
1341  
1342  	ret = dwxgmac2_filter_wait(hw);
1343  	if (ret)
1344  		return ret;
1345  
1346  	*data = readl(ioaddr + XGMAC_L3L4_DATA);
1347  	return 0;
1348  }
1349  
dwxgmac2_filter_write(struct mac_device_info * hw,u32 filter_no,u8 reg,u32 data)1350  static int dwxgmac2_filter_write(struct mac_device_info *hw, u32 filter_no,
1351  				 u8 reg, u32 data)
1352  {
1353  	void __iomem *ioaddr = hw->pcsr;
1354  	u32 value;
1355  	int ret;
1356  
1357  	ret = dwxgmac2_filter_wait(hw);
1358  	if (ret)
1359  		return ret;
1360  
1361  	writel(data, ioaddr + XGMAC_L3L4_DATA);
1362  
1363  	value = ((filter_no << XGMAC_IDDR_FNUM) | reg) << XGMAC_IDDR_SHIFT;
1364  	value |= XGMAC_XB;
1365  	writel(value, ioaddr + XGMAC_L3L4_ADDR_CTRL);
1366  
1367  	return dwxgmac2_filter_wait(hw);
1368  }
1369  
dwxgmac2_config_l3_filter(struct mac_device_info * hw,u32 filter_no,bool en,bool ipv6,bool sa,bool inv,u32 match)1370  static int dwxgmac2_config_l3_filter(struct mac_device_info *hw, u32 filter_no,
1371  				     bool en, bool ipv6, bool sa, bool inv,
1372  				     u32 match)
1373  {
1374  	void __iomem *ioaddr = hw->pcsr;
1375  	u32 value;
1376  	int ret;
1377  
1378  	value = readl(ioaddr + XGMAC_PACKET_FILTER);
1379  	value |= XGMAC_FILTER_IPFE;
1380  	writel(value, ioaddr + XGMAC_PACKET_FILTER);
1381  
1382  	ret = dwxgmac2_filter_read(hw, filter_no, XGMAC_L3L4_CTRL, &value);
1383  	if (ret)
1384  		return ret;
1385  
1386  	/* For IPv6 not both SA/DA filters can be active */
1387  	if (ipv6) {
1388  		value |= XGMAC_L3PEN0;
1389  		value &= ~(XGMAC_L3SAM0 | XGMAC_L3SAIM0);
1390  		value &= ~(XGMAC_L3DAM0 | XGMAC_L3DAIM0);
1391  		if (sa) {
1392  			value |= XGMAC_L3SAM0;
1393  			if (inv)
1394  				value |= XGMAC_L3SAIM0;
1395  		} else {
1396  			value |= XGMAC_L3DAM0;
1397  			if (inv)
1398  				value |= XGMAC_L3DAIM0;
1399  		}
1400  	} else {
1401  		value &= ~XGMAC_L3PEN0;
1402  		if (sa) {
1403  			value |= XGMAC_L3SAM0;
1404  			if (inv)
1405  				value |= XGMAC_L3SAIM0;
1406  		} else {
1407  			value |= XGMAC_L3DAM0;
1408  			if (inv)
1409  				value |= XGMAC_L3DAIM0;
1410  		}
1411  	}
1412  
1413  	ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, value);
1414  	if (ret)
1415  		return ret;
1416  
1417  	if (sa) {
1418  		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3_ADDR0, match);
1419  		if (ret)
1420  			return ret;
1421  	} else {
1422  		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3_ADDR1, match);
1423  		if (ret)
1424  			return ret;
1425  	}
1426  
1427  	if (!en)
1428  		return dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, 0);
1429  
1430  	return 0;
1431  }
1432  
dwxgmac2_config_l4_filter(struct mac_device_info * hw,u32 filter_no,bool en,bool udp,bool sa,bool inv,u32 match)1433  static int dwxgmac2_config_l4_filter(struct mac_device_info *hw, u32 filter_no,
1434  				     bool en, bool udp, bool sa, bool inv,
1435  				     u32 match)
1436  {
1437  	void __iomem *ioaddr = hw->pcsr;
1438  	u32 value;
1439  	int ret;
1440  
1441  	value = readl(ioaddr + XGMAC_PACKET_FILTER);
1442  	value |= XGMAC_FILTER_IPFE;
1443  	writel(value, ioaddr + XGMAC_PACKET_FILTER);
1444  
1445  	ret = dwxgmac2_filter_read(hw, filter_no, XGMAC_L3L4_CTRL, &value);
1446  	if (ret)
1447  		return ret;
1448  
1449  	if (udp) {
1450  		value |= XGMAC_L4PEN0;
1451  	} else {
1452  		value &= ~XGMAC_L4PEN0;
1453  	}
1454  
1455  	value &= ~(XGMAC_L4SPM0 | XGMAC_L4SPIM0);
1456  	value &= ~(XGMAC_L4DPM0 | XGMAC_L4DPIM0);
1457  	if (sa) {
1458  		value |= XGMAC_L4SPM0;
1459  		if (inv)
1460  			value |= XGMAC_L4SPIM0;
1461  	} else {
1462  		value |= XGMAC_L4DPM0;
1463  		if (inv)
1464  			value |= XGMAC_L4DPIM0;
1465  	}
1466  
1467  	ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, value);
1468  	if (ret)
1469  		return ret;
1470  
1471  	if (sa) {
1472  		value = match & XGMAC_L4SP0;
1473  
1474  		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L4_ADDR, value);
1475  		if (ret)
1476  			return ret;
1477  	} else {
1478  		value = (match << XGMAC_L4DP0_SHIFT) & XGMAC_L4DP0;
1479  
1480  		ret = dwxgmac2_filter_write(hw, filter_no, XGMAC_L4_ADDR, value);
1481  		if (ret)
1482  			return ret;
1483  	}
1484  
1485  	if (!en)
1486  		return dwxgmac2_filter_write(hw, filter_no, XGMAC_L3L4_CTRL, 0);
1487  
1488  	return 0;
1489  }
1490  
dwxgmac2_set_arp_offload(struct mac_device_info * hw,bool en,u32 addr)1491  static void dwxgmac2_set_arp_offload(struct mac_device_info *hw, bool en,
1492  				     u32 addr)
1493  {
1494  	void __iomem *ioaddr = hw->pcsr;
1495  	u32 value;
1496  
1497  	writel(addr, ioaddr + XGMAC_ARP_ADDR);
1498  
1499  	value = readl(ioaddr + XGMAC_RX_CONFIG);
1500  	if (en)
1501  		value |= XGMAC_CONFIG_ARPEN;
1502  	else
1503  		value &= ~XGMAC_CONFIG_ARPEN;
1504  	writel(value, ioaddr + XGMAC_RX_CONFIG);
1505  }
1506  
dwxgmac3_fpe_configure(void __iomem * ioaddr,struct stmmac_fpe_cfg * cfg,u32 num_txq,u32 num_rxq,bool tx_enable,bool pmac_enable)1507  static void dwxgmac3_fpe_configure(void __iomem *ioaddr,
1508  				   struct stmmac_fpe_cfg *cfg,
1509  				   u32 num_txq, u32 num_rxq,
1510  				   bool tx_enable, bool pmac_enable)
1511  {
1512  	u32 value;
1513  
1514  	if (!tx_enable) {
1515  		value = readl(ioaddr + XGMAC_FPE_CTRL_STS);
1516  
1517  		value &= ~XGMAC_EFPE;
1518  
1519  		writel(value, ioaddr + XGMAC_FPE_CTRL_STS);
1520  		return;
1521  	}
1522  
1523  	value = readl(ioaddr + XGMAC_RXQ_CTRL1);
1524  	value &= ~XGMAC_RQ;
1525  	value |= (num_rxq - 1) << XGMAC_RQ_SHIFT;
1526  	writel(value, ioaddr + XGMAC_RXQ_CTRL1);
1527  
1528  	value = readl(ioaddr + XGMAC_FPE_CTRL_STS);
1529  	value |= XGMAC_EFPE;
1530  	writel(value, ioaddr + XGMAC_FPE_CTRL_STS);
1531  }
1532  
1533  const struct stmmac_ops dwxgmac210_ops = {
1534  	.core_init = dwxgmac2_core_init,
1535  	.set_mac = dwxgmac2_set_mac,
1536  	.rx_ipc = dwxgmac2_rx_ipc,
1537  	.rx_queue_enable = dwxgmac2_rx_queue_enable,
1538  	.rx_queue_prio = dwxgmac2_rx_queue_prio,
1539  	.tx_queue_prio = dwxgmac2_tx_queue_prio,
1540  	.rx_queue_routing = dwxgmac2_rx_queue_routing,
1541  	.prog_mtl_rx_algorithms = dwxgmac2_prog_mtl_rx_algorithms,
1542  	.prog_mtl_tx_algorithms = dwxgmac2_prog_mtl_tx_algorithms,
1543  	.set_mtl_tx_queue_weight = dwxgmac2_set_mtl_tx_queue_weight,
1544  	.map_mtl_to_dma = dwxgmac2_map_mtl_to_dma,
1545  	.config_cbs = dwxgmac2_config_cbs,
1546  	.dump_regs = dwxgmac2_dump_regs,
1547  	.host_irq_status = dwxgmac2_host_irq_status,
1548  	.host_mtl_irq_status = dwxgmac2_host_mtl_irq_status,
1549  	.flow_ctrl = dwxgmac2_flow_ctrl,
1550  	.pmt = dwxgmac2_pmt,
1551  	.set_umac_addr = dwxgmac2_set_umac_addr,
1552  	.get_umac_addr = dwxgmac2_get_umac_addr,
1553  	.set_eee_mode = dwxgmac2_set_eee_mode,
1554  	.reset_eee_mode = dwxgmac2_reset_eee_mode,
1555  	.set_eee_timer = dwxgmac2_set_eee_timer,
1556  	.set_eee_pls = dwxgmac2_set_eee_pls,
1557  	.debug = NULL,
1558  	.set_filter = dwxgmac2_set_filter,
1559  	.safety_feat_config = dwxgmac3_safety_feat_config,
1560  	.safety_feat_irq_status = dwxgmac3_safety_feat_irq_status,
1561  	.safety_feat_dump = dwxgmac3_safety_feat_dump,
1562  	.set_mac_loopback = dwxgmac2_set_mac_loopback,
1563  	.rss_configure = dwxgmac2_rss_configure,
1564  	.update_vlan_hash = dwxgmac2_update_vlan_hash,
1565  	.rxp_config = dwxgmac3_rxp_config,
1566  	.get_mac_tx_timestamp = dwxgmac2_get_mac_tx_timestamp,
1567  	.flex_pps_config = dwxgmac2_flex_pps_config,
1568  	.sarc_configure = dwxgmac2_sarc_configure,
1569  	.enable_vlan = dwxgmac2_enable_vlan,
1570  	.config_l3_filter = dwxgmac2_config_l3_filter,
1571  	.config_l4_filter = dwxgmac2_config_l4_filter,
1572  	.set_arp_offload = dwxgmac2_set_arp_offload,
1573  	.fpe_configure = dwxgmac3_fpe_configure,
1574  };
1575  
dwxlgmac2_rx_queue_enable(struct mac_device_info * hw,u8 mode,u32 queue)1576  static void dwxlgmac2_rx_queue_enable(struct mac_device_info *hw, u8 mode,
1577  				      u32 queue)
1578  {
1579  	void __iomem *ioaddr = hw->pcsr;
1580  	u32 value;
1581  
1582  	value = readl(ioaddr + XLGMAC_RXQ_ENABLE_CTRL0) & ~XGMAC_RXQEN(queue);
1583  	if (mode == MTL_QUEUE_AVB)
1584  		value |= 0x1 << XGMAC_RXQEN_SHIFT(queue);
1585  	else if (mode == MTL_QUEUE_DCB)
1586  		value |= 0x2 << XGMAC_RXQEN_SHIFT(queue);
1587  	writel(value, ioaddr + XLGMAC_RXQ_ENABLE_CTRL0);
1588  }
1589  
1590  const struct stmmac_ops dwxlgmac2_ops = {
1591  	.core_init = dwxgmac2_core_init,
1592  	.set_mac = dwxgmac2_set_mac,
1593  	.rx_ipc = dwxgmac2_rx_ipc,
1594  	.rx_queue_enable = dwxlgmac2_rx_queue_enable,
1595  	.rx_queue_prio = dwxgmac2_rx_queue_prio,
1596  	.tx_queue_prio = dwxgmac2_tx_queue_prio,
1597  	.rx_queue_routing = dwxgmac2_rx_queue_routing,
1598  	.prog_mtl_rx_algorithms = dwxgmac2_prog_mtl_rx_algorithms,
1599  	.prog_mtl_tx_algorithms = dwxgmac2_prog_mtl_tx_algorithms,
1600  	.set_mtl_tx_queue_weight = dwxgmac2_set_mtl_tx_queue_weight,
1601  	.map_mtl_to_dma = dwxgmac2_map_mtl_to_dma,
1602  	.config_cbs = dwxgmac2_config_cbs,
1603  	.dump_regs = dwxgmac2_dump_regs,
1604  	.host_irq_status = dwxgmac2_host_irq_status,
1605  	.host_mtl_irq_status = dwxgmac2_host_mtl_irq_status,
1606  	.flow_ctrl = dwxgmac2_flow_ctrl,
1607  	.pmt = dwxgmac2_pmt,
1608  	.set_umac_addr = dwxgmac2_set_umac_addr,
1609  	.get_umac_addr = dwxgmac2_get_umac_addr,
1610  	.set_eee_mode = dwxgmac2_set_eee_mode,
1611  	.reset_eee_mode = dwxgmac2_reset_eee_mode,
1612  	.set_eee_timer = dwxgmac2_set_eee_timer,
1613  	.set_eee_pls = dwxgmac2_set_eee_pls,
1614  	.debug = NULL,
1615  	.set_filter = dwxgmac2_set_filter,
1616  	.safety_feat_config = dwxgmac3_safety_feat_config,
1617  	.safety_feat_irq_status = dwxgmac3_safety_feat_irq_status,
1618  	.safety_feat_dump = dwxgmac3_safety_feat_dump,
1619  	.set_mac_loopback = dwxgmac2_set_mac_loopback,
1620  	.rss_configure = dwxgmac2_rss_configure,
1621  	.update_vlan_hash = dwxgmac2_update_vlan_hash,
1622  	.rxp_config = dwxgmac3_rxp_config,
1623  	.get_mac_tx_timestamp = dwxgmac2_get_mac_tx_timestamp,
1624  	.flex_pps_config = dwxgmac2_flex_pps_config,
1625  	.sarc_configure = dwxgmac2_sarc_configure,
1626  	.enable_vlan = dwxgmac2_enable_vlan,
1627  	.config_l3_filter = dwxgmac2_config_l3_filter,
1628  	.config_l4_filter = dwxgmac2_config_l4_filter,
1629  	.set_arp_offload = dwxgmac2_set_arp_offload,
1630  	.fpe_configure = dwxgmac3_fpe_configure,
1631  };
1632  
dwxgmac2_setup(struct stmmac_priv * priv)1633  int dwxgmac2_setup(struct stmmac_priv *priv)
1634  {
1635  	struct mac_device_info *mac = priv->hw;
1636  
1637  	dev_info(priv->device, "\tXGMAC2\n");
1638  
1639  	priv->dev->priv_flags |= IFF_UNICAST_FLT;
1640  	mac->pcsr = priv->ioaddr;
1641  	mac->multicast_filter_bins = priv->plat->multicast_filter_bins;
1642  	mac->unicast_filter_entries = priv->plat->unicast_filter_entries;
1643  	mac->mcast_bits_log2 = 0;
1644  
1645  	if (mac->multicast_filter_bins)
1646  		mac->mcast_bits_log2 = ilog2(mac->multicast_filter_bins);
1647  
1648  	mac->link.caps = MAC_ASYM_PAUSE | MAC_SYM_PAUSE |
1649  			 MAC_1000FD | MAC_2500FD | MAC_5000FD |
1650  			 MAC_10000FD;
1651  	mac->link.duplex = 0;
1652  	mac->link.speed10 = XGMAC_CONFIG_SS_10_MII;
1653  	mac->link.speed100 = XGMAC_CONFIG_SS_100_MII;
1654  	mac->link.speed1000 = XGMAC_CONFIG_SS_1000_GMII;
1655  	mac->link.speed2500 = XGMAC_CONFIG_SS_2500_GMII;
1656  	mac->link.xgmii.speed2500 = XGMAC_CONFIG_SS_2500;
1657  	mac->link.xgmii.speed5000 = XGMAC_CONFIG_SS_5000;
1658  	mac->link.xgmii.speed10000 = XGMAC_CONFIG_SS_10000;
1659  	mac->link.speed_mask = XGMAC_CONFIG_SS_MASK;
1660  
1661  	mac->mii.addr = XGMAC_MDIO_ADDR;
1662  	mac->mii.data = XGMAC_MDIO_DATA;
1663  	mac->mii.addr_shift = 16;
1664  	mac->mii.addr_mask = GENMASK(20, 16);
1665  	mac->mii.reg_shift = 0;
1666  	mac->mii.reg_mask = GENMASK(15, 0);
1667  	mac->mii.clk_csr_shift = 19;
1668  	mac->mii.clk_csr_mask = GENMASK(21, 19);
1669  
1670  	return 0;
1671  }
1672  
dwxlgmac2_setup(struct stmmac_priv * priv)1673  int dwxlgmac2_setup(struct stmmac_priv *priv)
1674  {
1675  	struct mac_device_info *mac = priv->hw;
1676  
1677  	dev_info(priv->device, "\tXLGMAC\n");
1678  
1679  	priv->dev->priv_flags |= IFF_UNICAST_FLT;
1680  	mac->pcsr = priv->ioaddr;
1681  	mac->multicast_filter_bins = priv->plat->multicast_filter_bins;
1682  	mac->unicast_filter_entries = priv->plat->unicast_filter_entries;
1683  	mac->mcast_bits_log2 = 0;
1684  
1685  	if (mac->multicast_filter_bins)
1686  		mac->mcast_bits_log2 = ilog2(mac->multicast_filter_bins);
1687  
1688  	mac->link.caps = MAC_ASYM_PAUSE | MAC_SYM_PAUSE |
1689  			 MAC_1000FD | MAC_2500FD | MAC_5000FD |
1690  			 MAC_10000FD | MAC_25000FD |
1691  			 MAC_40000FD | MAC_50000FD |
1692  			 MAC_100000FD;
1693  	mac->link.duplex = 0;
1694  	mac->link.speed1000 = XLGMAC_CONFIG_SS_1000;
1695  	mac->link.speed2500 = XLGMAC_CONFIG_SS_2500;
1696  	mac->link.xgmii.speed10000 = XLGMAC_CONFIG_SS_10G;
1697  	mac->link.xlgmii.speed25000 = XLGMAC_CONFIG_SS_25G;
1698  	mac->link.xlgmii.speed40000 = XLGMAC_CONFIG_SS_40G;
1699  	mac->link.xlgmii.speed50000 = XLGMAC_CONFIG_SS_50G;
1700  	mac->link.xlgmii.speed100000 = XLGMAC_CONFIG_SS_100G;
1701  	mac->link.speed_mask = XLGMAC_CONFIG_SS;
1702  
1703  	mac->mii.addr = XGMAC_MDIO_ADDR;
1704  	mac->mii.data = XGMAC_MDIO_DATA;
1705  	mac->mii.addr_shift = 16;
1706  	mac->mii.addr_mask = GENMASK(20, 16);
1707  	mac->mii.reg_shift = 0;
1708  	mac->mii.reg_mask = GENMASK(15, 0);
1709  	mac->mii.clk_csr_shift = 19;
1710  	mac->mii.clk_csr_mask = GENMASK(21, 19);
1711  
1712  	return 0;
1713  }
1714