treewide: Use fallthrough pseudo-keyword
[linux-block.git] / drivers / net / ethernet / ibm / emac / core.c
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  * drivers/net/ethernet/ibm/emac/core.c
4  *
5  * Driver for PowerPC 4xx on-chip ethernet controller.
6  *
7  * Copyright 2007 Benjamin Herrenschmidt, IBM Corp.
8  *                <benh@kernel.crashing.org>
9  *
10  * Based on the arch/ppc version of the driver:
11  *
12  * Copyright (c) 2004, 2005 Zultys Technologies.
13  * Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>
14  *
15  * Based on original work by
16  *      Matt Porter <mporter@kernel.crashing.org>
17  *      (c) 2003 Benjamin Herrenschmidt <benh@kernel.crashing.org>
18  *      Armin Kuster <akuster@mvista.com>
19  *      Johnnie Peters <jpeters@mvista.com>
20  */
21
22 #include <linux/module.h>
23 #include <linux/sched.h>
24 #include <linux/string.h>
25 #include <linux/errno.h>
26 #include <linux/delay.h>
27 #include <linux/types.h>
28 #include <linux/pci.h>
29 #include <linux/etherdevice.h>
30 #include <linux/skbuff.h>
31 #include <linux/crc32.h>
32 #include <linux/ethtool.h>
33 #include <linux/mii.h>
34 #include <linux/bitops.h>
35 #include <linux/workqueue.h>
36 #include <linux/of.h>
37 #include <linux/of_address.h>
38 #include <linux/of_irq.h>
39 #include <linux/of_net.h>
40 #include <linux/of_mdio.h>
41 #include <linux/slab.h>
42
43 #include <asm/processor.h>
44 #include <asm/io.h>
45 #include <asm/dma.h>
46 #include <linux/uaccess.h>
47 #include <asm/dcr.h>
48 #include <asm/dcr-regs.h>
49
50 #include "core.h"
51
52 /*
53  * Lack of dma_unmap_???? calls is intentional.
54  *
55  * API-correct usage requires additional support state information to be
56  * maintained for every RX and TX buffer descriptor (BD). Unfortunately, due to
57  * EMAC design (e.g. TX buffer passed from network stack can be split into
58  * several BDs, dma_map_single/dma_map_page can be used to map particular BD),
59  * maintaining such information will add additional overhead.
60  * Current DMA API implementation for 4xx processors only ensures cache coherency
61  * and dma_unmap_???? routines are empty and are likely to stay this way.
62  * I decided to omit dma_unmap_??? calls because I don't want to add additional
63  * complexity just for the sake of following some abstract API, when it doesn't
64  * add any real benefit to the driver. I understand that this decision maybe
65  * controversial, but I really tried to make code API-correct and efficient
66  * at the same time and didn't come up with code I liked :(.                --ebs
67  */
68
69 #define DRV_NAME        "emac"
70 #define DRV_VERSION     "3.54"
71 #define DRV_DESC        "PPC 4xx OCP EMAC driver"
72
73 MODULE_DESCRIPTION(DRV_DESC);
74 MODULE_AUTHOR
75     ("Eugene Surovegin <eugene.surovegin@zultys.com> or <ebs@ebshome.net>");
76 MODULE_LICENSE("GPL");
77
78 /* minimum number of free TX descriptors required to wake up TX process */
79 #define EMAC_TX_WAKEUP_THRESH           (NUM_TX_BUFF / 4)
80
81 /* If packet size is less than this number, we allocate small skb and copy packet
82  * contents into it instead of just sending original big skb up
83  */
84 #define EMAC_RX_COPY_THRESH             CONFIG_IBM_EMAC_RX_COPY_THRESHOLD
85
86 /* Since multiple EMACs share MDIO lines in various ways, we need
87  * to avoid re-using the same PHY ID in cases where the arch didn't
88  * setup precise phy_map entries
89  *
90  * XXX This is something that needs to be reworked as we can have multiple
91  * EMAC "sets" (multiple ASICs containing several EMACs) though we can
92  * probably require in that case to have explicit PHY IDs in the device-tree
93  */
94 static u32 busy_phy_map;
95 static DEFINE_MUTEX(emac_phy_map_lock);
96
97 /* This is the wait queue used to wait on any event related to probe, that
98  * is discovery of MALs, other EMACs, ZMII/RGMIIs, etc...
99  */
100 static DECLARE_WAIT_QUEUE_HEAD(emac_probe_wait);
101
102 /* Having stable interface names is a doomed idea. However, it would be nice
103  * if we didn't have completely random interface names at boot too :-) It's
104  * just a matter of making everybody's life easier. Since we are doing
105  * threaded probing, it's a bit harder though. The base idea here is that
106  * we make up a list of all emacs in the device-tree before we register the
107  * driver. Every emac will then wait for the previous one in the list to
108  * initialize before itself. We should also keep that list ordered by
109  * cell_index.
110  * That list is only 4 entries long, meaning that additional EMACs don't
111  * get ordering guarantees unless EMAC_BOOT_LIST_SIZE is increased.
112  */
113
114 #define EMAC_BOOT_LIST_SIZE     4
115 static struct device_node *emac_boot_list[EMAC_BOOT_LIST_SIZE];
116
117 /* How long should I wait for dependent devices ? */
118 #define EMAC_PROBE_DEP_TIMEOUT  (HZ * 5)
119
120 /* I don't want to litter system log with timeout errors
121  * when we have brain-damaged PHY.
122  */
123 static inline void emac_report_timeout_error(struct emac_instance *dev,
124                                              const char *error)
125 {
126         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX |
127                                   EMAC_FTR_460EX_PHY_CLK_FIX |
128                                   EMAC_FTR_440EP_PHY_CLK_FIX))
129                 DBG(dev, "%s" NL, error);
130         else if (net_ratelimit())
131                 printk(KERN_ERR "%pOF: %s\n", dev->ofdev->dev.of_node, error);
132 }
133
134 /* EMAC PHY clock workaround:
135  * 440EP/440GR has more sane SDR0_MFR register implementation than 440GX,
136  * which allows controlling each EMAC clock
137  */
138 static inline void emac_rx_clk_tx(struct emac_instance *dev)
139 {
140 #ifdef CONFIG_PPC_DCR_NATIVE
141         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
142                 dcri_clrset(SDR0, SDR0_MFR,
143                             0, SDR0_MFR_ECS >> dev->cell_index);
144 #endif
145 }
146
147 static inline void emac_rx_clk_default(struct emac_instance *dev)
148 {
149 #ifdef CONFIG_PPC_DCR_NATIVE
150         if (emac_has_feature(dev, EMAC_FTR_440EP_PHY_CLK_FIX))
151                 dcri_clrset(SDR0, SDR0_MFR,
152                             SDR0_MFR_ECS >> dev->cell_index, 0);
153 #endif
154 }
155
156 /* PHY polling intervals */
157 #define PHY_POLL_LINK_ON        HZ
158 #define PHY_POLL_LINK_OFF       (HZ / 5)
159
160 /* Graceful stop timeouts in us.
161  * We should allow up to 1 frame time (full-duplex, ignoring collisions)
162  */
163 #define STOP_TIMEOUT_10         1230
164 #define STOP_TIMEOUT_100        124
165 #define STOP_TIMEOUT_1000       13
166 #define STOP_TIMEOUT_1000_JUMBO 73
167
168 static unsigned char default_mcast_addr[] = {
169         0x01, 0x80, 0xC2, 0x00, 0x00, 0x01
170 };
171
172 /* Please, keep in sync with struct ibm_emac_stats/ibm_emac_error_stats */
173 static const char emac_stats_keys[EMAC_ETHTOOL_STATS_COUNT][ETH_GSTRING_LEN] = {
174         "rx_packets", "rx_bytes", "tx_packets", "tx_bytes", "rx_packets_csum",
175         "tx_packets_csum", "tx_undo", "rx_dropped_stack", "rx_dropped_oom",
176         "rx_dropped_error", "rx_dropped_resize", "rx_dropped_mtu",
177         "rx_stopped", "rx_bd_errors", "rx_bd_overrun", "rx_bd_bad_packet",
178         "rx_bd_runt_packet", "rx_bd_short_event", "rx_bd_alignment_error",
179         "rx_bd_bad_fcs", "rx_bd_packet_too_long", "rx_bd_out_of_range",
180         "rx_bd_in_range", "rx_parity", "rx_fifo_overrun", "rx_overrun",
181         "rx_bad_packet", "rx_runt_packet", "rx_short_event",
182         "rx_alignment_error", "rx_bad_fcs", "rx_packet_too_long",
183         "rx_out_of_range", "rx_in_range", "tx_dropped", "tx_bd_errors",
184         "tx_bd_bad_fcs", "tx_bd_carrier_loss", "tx_bd_excessive_deferral",
185         "tx_bd_excessive_collisions", "tx_bd_late_collision",
186         "tx_bd_multple_collisions", "tx_bd_single_collision",
187         "tx_bd_underrun", "tx_bd_sqe", "tx_parity", "tx_underrun", "tx_sqe",
188         "tx_errors"
189 };
190
191 static irqreturn_t emac_irq(int irq, void *dev_instance);
192 static void emac_clean_tx_ring(struct emac_instance *dev);
193 static void __emac_set_multicast_list(struct emac_instance *dev);
194
195 static inline int emac_phy_supports_gige(int phy_mode)
196 {
197         return  phy_interface_mode_is_rgmii(phy_mode) ||
198                 phy_mode == PHY_INTERFACE_MODE_GMII ||
199                 phy_mode == PHY_INTERFACE_MODE_SGMII ||
200                 phy_mode == PHY_INTERFACE_MODE_TBI ||
201                 phy_mode == PHY_INTERFACE_MODE_RTBI;
202 }
203
204 static inline int emac_phy_gpcs(int phy_mode)
205 {
206         return  phy_mode == PHY_INTERFACE_MODE_SGMII ||
207                 phy_mode == PHY_INTERFACE_MODE_TBI ||
208                 phy_mode == PHY_INTERFACE_MODE_RTBI;
209 }
210
211 static inline void emac_tx_enable(struct emac_instance *dev)
212 {
213         struct emac_regs __iomem *p = dev->emacp;
214         u32 r;
215
216         DBG(dev, "tx_enable" NL);
217
218         r = in_be32(&p->mr0);
219         if (!(r & EMAC_MR0_TXE))
220                 out_be32(&p->mr0, r | EMAC_MR0_TXE);
221 }
222
223 static void emac_tx_disable(struct emac_instance *dev)
224 {
225         struct emac_regs __iomem *p = dev->emacp;
226         u32 r;
227
228         DBG(dev, "tx_disable" NL);
229
230         r = in_be32(&p->mr0);
231         if (r & EMAC_MR0_TXE) {
232                 int n = dev->stop_timeout;
233                 out_be32(&p->mr0, r & ~EMAC_MR0_TXE);
234                 while (!(in_be32(&p->mr0) & EMAC_MR0_TXI) && n) {
235                         udelay(1);
236                         --n;
237                 }
238                 if (unlikely(!n))
239                         emac_report_timeout_error(dev, "TX disable timeout");
240         }
241 }
242
243 static void emac_rx_enable(struct emac_instance *dev)
244 {
245         struct emac_regs __iomem *p = dev->emacp;
246         u32 r;
247
248         if (unlikely(test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags)))
249                 goto out;
250
251         DBG(dev, "rx_enable" NL);
252
253         r = in_be32(&p->mr0);
254         if (!(r & EMAC_MR0_RXE)) {
255                 if (unlikely(!(r & EMAC_MR0_RXI))) {
256                         /* Wait if previous async disable is still in progress */
257                         int n = dev->stop_timeout;
258                         while (!(r = in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
259                                 udelay(1);
260                                 --n;
261                         }
262                         if (unlikely(!n))
263                                 emac_report_timeout_error(dev,
264                                                           "RX disable timeout");
265                 }
266                 out_be32(&p->mr0, r | EMAC_MR0_RXE);
267         }
268  out:
269         ;
270 }
271
272 static void emac_rx_disable(struct emac_instance *dev)
273 {
274         struct emac_regs __iomem *p = dev->emacp;
275         u32 r;
276
277         DBG(dev, "rx_disable" NL);
278
279         r = in_be32(&p->mr0);
280         if (r & EMAC_MR0_RXE) {
281                 int n = dev->stop_timeout;
282                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
283                 while (!(in_be32(&p->mr0) & EMAC_MR0_RXI) && n) {
284                         udelay(1);
285                         --n;
286                 }
287                 if (unlikely(!n))
288                         emac_report_timeout_error(dev, "RX disable timeout");
289         }
290 }
291
292 static inline void emac_netif_stop(struct emac_instance *dev)
293 {
294         netif_tx_lock_bh(dev->ndev);
295         netif_addr_lock(dev->ndev);
296         dev->no_mcast = 1;
297         netif_addr_unlock(dev->ndev);
298         netif_tx_unlock_bh(dev->ndev);
299         netif_trans_update(dev->ndev);  /* prevent tx timeout */
300         mal_poll_disable(dev->mal, &dev->commac);
301         netif_tx_disable(dev->ndev);
302 }
303
304 static inline void emac_netif_start(struct emac_instance *dev)
305 {
306         netif_tx_lock_bh(dev->ndev);
307         netif_addr_lock(dev->ndev);
308         dev->no_mcast = 0;
309         if (dev->mcast_pending && netif_running(dev->ndev))
310                 __emac_set_multicast_list(dev);
311         netif_addr_unlock(dev->ndev);
312         netif_tx_unlock_bh(dev->ndev);
313
314         netif_wake_queue(dev->ndev);
315
316         /* NOTE: unconditional netif_wake_queue is only appropriate
317          * so long as all callers are assured to have free tx slots
318          * (taken from tg3... though the case where that is wrong is
319          *  not terribly harmful)
320          */
321         mal_poll_enable(dev->mal, &dev->commac);
322 }
323
324 static inline void emac_rx_disable_async(struct emac_instance *dev)
325 {
326         struct emac_regs __iomem *p = dev->emacp;
327         u32 r;
328
329         DBG(dev, "rx_disable_async" NL);
330
331         r = in_be32(&p->mr0);
332         if (r & EMAC_MR0_RXE)
333                 out_be32(&p->mr0, r & ~EMAC_MR0_RXE);
334 }
335
336 static int emac_reset(struct emac_instance *dev)
337 {
338         struct emac_regs __iomem *p = dev->emacp;
339         int n = 20;
340         bool __maybe_unused try_internal_clock = false;
341
342         DBG(dev, "reset" NL);
343
344         if (!dev->reset_failed) {
345                 /* 40x erratum suggests stopping RX channel before reset,
346                  * we stop TX as well
347                  */
348                 emac_rx_disable(dev);
349                 emac_tx_disable(dev);
350         }
351
352 #ifdef CONFIG_PPC_DCR_NATIVE
353 do_retry:
354         /*
355          * PPC460EX/GT Embedded Processor Advanced User's Manual
356          * section 28.10.1 Mode Register 0 (EMACx_MR0) states:
357          * Note: The PHY must provide a TX Clk in order to perform a soft reset
358          * of the EMAC. If none is present, select the internal clock
359          * (SDR0_ETH_CFG[EMACx_PHY_CLK] = 1).
360          * After a soft reset, select the external clock.
361          *
362          * The AR8035-A PHY Meraki MR24 does not provide a TX Clk if the
363          * ethernet cable is not attached. This causes the reset to timeout
364          * and the PHY detection code in emac_init_phy() is unable to
365          * communicate and detect the AR8035-A PHY. As a result, the emac
366          * driver bails out early and the user has no ethernet.
367          * In order to stay compatible with existing configurations, the
368          * driver will temporarily switch to the internal clock, after
369          * the first reset fails.
370          */
371         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX)) {
372                 if (try_internal_clock || (dev->phy_address == 0xffffffff &&
373                                            dev->phy_map == 0xffffffff)) {
374                         /* No PHY: select internal loop clock before reset */
375                         dcri_clrset(SDR0, SDR0_ETH_CFG,
376                                     0, SDR0_ETH_CFG_ECS << dev->cell_index);
377                 } else {
378                         /* PHY present: select external clock before reset */
379                         dcri_clrset(SDR0, SDR0_ETH_CFG,
380                                     SDR0_ETH_CFG_ECS << dev->cell_index, 0);
381                 }
382         }
383 #endif
384
385         out_be32(&p->mr0, EMAC_MR0_SRST);
386         while ((in_be32(&p->mr0) & EMAC_MR0_SRST) && n)
387                 --n;
388
389 #ifdef CONFIG_PPC_DCR_NATIVE
390         if (emac_has_feature(dev, EMAC_FTR_460EX_PHY_CLK_FIX)) {
391                 if (!n && !try_internal_clock) {
392                         /* first attempt has timed out. */
393                         n = 20;
394                         try_internal_clock = true;
395                         goto do_retry;
396                 }
397
398                 if (try_internal_clock || (dev->phy_address == 0xffffffff &&
399                                            dev->phy_map == 0xffffffff)) {
400                         /* No PHY: restore external clock source after reset */
401                         dcri_clrset(SDR0, SDR0_ETH_CFG,
402                                     SDR0_ETH_CFG_ECS << dev->cell_index, 0);
403                 }
404         }
405 #endif
406
407         if (n) {
408                 dev->reset_failed = 0;
409                 return 0;
410         } else {
411                 emac_report_timeout_error(dev, "reset timeout");
412                 dev->reset_failed = 1;
413                 return -ETIMEDOUT;
414         }
415 }
416
417 static void emac_hash_mc(struct emac_instance *dev)
418 {
419         const int regs = EMAC_XAHT_REGS(dev);
420         u32 *gaht_base = emac_gaht_base(dev);
421         u32 gaht_temp[EMAC_XAHT_MAX_REGS];
422         struct netdev_hw_addr *ha;
423         int i;
424
425         DBG(dev, "hash_mc %d" NL, netdev_mc_count(dev->ndev));
426
427         memset(gaht_temp, 0, sizeof (gaht_temp));
428
429         netdev_for_each_mc_addr(ha, dev->ndev) {
430                 int slot, reg, mask;
431                 DBG2(dev, "mc %pM" NL, ha->addr);
432
433                 slot = EMAC_XAHT_CRC_TO_SLOT(dev,
434                                              ether_crc(ETH_ALEN, ha->addr));
435                 reg = EMAC_XAHT_SLOT_TO_REG(dev, slot);
436                 mask = EMAC_XAHT_SLOT_TO_MASK(dev, slot);
437
438                 gaht_temp[reg] |= mask;
439         }
440
441         for (i = 0; i < regs; i++)
442                 out_be32(gaht_base + i, gaht_temp[i]);
443 }
444
445 static inline u32 emac_iff2rmr(struct net_device *ndev)
446 {
447         struct emac_instance *dev = netdev_priv(ndev);
448         u32 r;
449
450         r = EMAC_RMR_SP | EMAC_RMR_SFCS | EMAC_RMR_IAE | EMAC_RMR_BAE;
451
452         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
453             r |= EMAC4_RMR_BASE;
454         else
455             r |= EMAC_RMR_BASE;
456
457         if (ndev->flags & IFF_PROMISC)
458                 r |= EMAC_RMR_PME;
459         else if (ndev->flags & IFF_ALLMULTI ||
460                          (netdev_mc_count(ndev) > EMAC_XAHT_SLOTS(dev)))
461                 r |= EMAC_RMR_PMME;
462         else if (!netdev_mc_empty(ndev))
463                 r |= EMAC_RMR_MAE;
464
465         if (emac_has_feature(dev, EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE)) {
466                 r &= ~EMAC4_RMR_MJS_MASK;
467                 r |= EMAC4_RMR_MJS(ndev->mtu);
468         }
469
470         return r;
471 }
472
473 static u32 __emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
474 {
475         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC_MR1_TR0_MULT;
476
477         DBG2(dev, "__emac_calc_base_mr1" NL);
478
479         switch(tx_size) {
480         case 2048:
481                 ret |= EMAC_MR1_TFS_2K;
482                 break;
483         default:
484                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
485                        dev->ndev->name, tx_size);
486         }
487
488         switch(rx_size) {
489         case 16384:
490                 ret |= EMAC_MR1_RFS_16K;
491                 break;
492         case 4096:
493                 ret |= EMAC_MR1_RFS_4K;
494                 break;
495         default:
496                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
497                        dev->ndev->name, rx_size);
498         }
499
500         return ret;
501 }
502
503 static u32 __emac4_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
504 {
505         u32 ret = EMAC_MR1_VLE | EMAC_MR1_IST | EMAC4_MR1_TR |
506                 EMAC4_MR1_OBCI(dev->opb_bus_freq / 1000000);
507
508         DBG2(dev, "__emac4_calc_base_mr1" NL);
509
510         switch(tx_size) {
511         case 16384:
512                 ret |= EMAC4_MR1_TFS_16K;
513                 break;
514         case 8192:
515                 ret |= EMAC4_MR1_TFS_8K;
516                 break;
517         case 4096:
518                 ret |= EMAC4_MR1_TFS_4K;
519                 break;
520         case 2048:
521                 ret |= EMAC4_MR1_TFS_2K;
522                 break;
523         default:
524                 printk(KERN_WARNING "%s: Unknown Tx FIFO size %d\n",
525                        dev->ndev->name, tx_size);
526         }
527
528         switch(rx_size) {
529         case 16384:
530                 ret |= EMAC4_MR1_RFS_16K;
531                 break;
532         case 8192:
533                 ret |= EMAC4_MR1_RFS_8K;
534                 break;
535         case 4096:
536                 ret |= EMAC4_MR1_RFS_4K;
537                 break;
538         case 2048:
539                 ret |= EMAC4_MR1_RFS_2K;
540                 break;
541         default:
542                 printk(KERN_WARNING "%s: Unknown Rx FIFO size %d\n",
543                        dev->ndev->name, rx_size);
544         }
545
546         return ret;
547 }
548
549 static u32 emac_calc_base_mr1(struct emac_instance *dev, int tx_size, int rx_size)
550 {
551         return emac_has_feature(dev, EMAC_FTR_EMAC4) ?
552                 __emac4_calc_base_mr1(dev, tx_size, rx_size) :
553                 __emac_calc_base_mr1(dev, tx_size, rx_size);
554 }
555
556 static inline u32 emac_calc_trtr(struct emac_instance *dev, unsigned int size)
557 {
558         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
559                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT_EMAC4;
560         else
561                 return ((size >> 6) - 1) << EMAC_TRTR_SHIFT;
562 }
563
564 static inline u32 emac_calc_rwmr(struct emac_instance *dev,
565                                  unsigned int low, unsigned int high)
566 {
567         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
568                 return (low << 22) | ( (high & 0x3ff) << 6);
569         else
570                 return (low << 23) | ( (high & 0x1ff) << 7);
571 }
572
573 static int emac_configure(struct emac_instance *dev)
574 {
575         struct emac_regs __iomem *p = dev->emacp;
576         struct net_device *ndev = dev->ndev;
577         int tx_size, rx_size, link = netif_carrier_ok(dev->ndev);
578         u32 r, mr1 = 0;
579
580         DBG(dev, "configure" NL);
581
582         if (!link) {
583                 out_be32(&p->mr1, in_be32(&p->mr1)
584                          | EMAC_MR1_FDE | EMAC_MR1_ILE);
585                 udelay(100);
586         } else if (emac_reset(dev) < 0)
587                 return -ETIMEDOUT;
588
589         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
590                 tah_reset(dev->tah_dev);
591
592         DBG(dev, " link = %d duplex = %d, pause = %d, asym_pause = %d\n",
593             link, dev->phy.duplex, dev->phy.pause, dev->phy.asym_pause);
594
595         /* Default fifo sizes */
596         tx_size = dev->tx_fifo_size;
597         rx_size = dev->rx_fifo_size;
598
599         /* No link, force loopback */
600         if (!link)
601                 mr1 = EMAC_MR1_FDE | EMAC_MR1_ILE;
602
603         /* Check for full duplex */
604         else if (dev->phy.duplex == DUPLEX_FULL)
605                 mr1 |= EMAC_MR1_FDE | EMAC_MR1_MWSW_001;
606
607         /* Adjust fifo sizes, mr1 and timeouts based on link speed */
608         dev->stop_timeout = STOP_TIMEOUT_10;
609         switch (dev->phy.speed) {
610         case SPEED_1000:
611                 if (emac_phy_gpcs(dev->phy.mode)) {
612                         mr1 |= EMAC_MR1_MF_1000GPCS | EMAC_MR1_MF_IPPA(
613                                 (dev->phy.gpcs_address != 0xffffffff) ?
614                                  dev->phy.gpcs_address : dev->phy.address);
615
616                         /* Put some arbitrary OUI, Manuf & Rev IDs so we can
617                          * identify this GPCS PHY later.
618                          */
619                         out_be32(&p->u1.emac4.ipcr, 0xdeadbeef);
620                 } else
621                         mr1 |= EMAC_MR1_MF_1000;
622
623                 /* Extended fifo sizes */
624                 tx_size = dev->tx_fifo_size_gige;
625                 rx_size = dev->rx_fifo_size_gige;
626
627                 if (dev->ndev->mtu > ETH_DATA_LEN) {
628                         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
629                                 mr1 |= EMAC4_MR1_JPSM;
630                         else
631                                 mr1 |= EMAC_MR1_JPSM;
632                         dev->stop_timeout = STOP_TIMEOUT_1000_JUMBO;
633                 } else
634                         dev->stop_timeout = STOP_TIMEOUT_1000;
635                 break;
636         case SPEED_100:
637                 mr1 |= EMAC_MR1_MF_100;
638                 dev->stop_timeout = STOP_TIMEOUT_100;
639                 break;
640         default: /* make gcc happy */
641                 break;
642         }
643
644         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
645                 rgmii_set_speed(dev->rgmii_dev, dev->rgmii_port,
646                                 dev->phy.speed);
647         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
648                 zmii_set_speed(dev->zmii_dev, dev->zmii_port, dev->phy.speed);
649
650         /* on 40x erratum forces us to NOT use integrated flow control,
651          * let's hope it works on 44x ;)
652          */
653         if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x) &&
654             dev->phy.duplex == DUPLEX_FULL) {
655                 if (dev->phy.pause)
656                         mr1 |= EMAC_MR1_EIFC | EMAC_MR1_APP;
657                 else if (dev->phy.asym_pause)
658                         mr1 |= EMAC_MR1_APP;
659         }
660
661         /* Add base settings & fifo sizes & program MR1 */
662         mr1 |= emac_calc_base_mr1(dev, tx_size, rx_size);
663         out_be32(&p->mr1, mr1);
664
665         /* Set individual MAC address */
666         out_be32(&p->iahr, (ndev->dev_addr[0] << 8) | ndev->dev_addr[1]);
667         out_be32(&p->ialr, (ndev->dev_addr[2] << 24) |
668                  (ndev->dev_addr[3] << 16) | (ndev->dev_addr[4] << 8) |
669                  ndev->dev_addr[5]);
670
671         /* VLAN Tag Protocol ID */
672         out_be32(&p->vtpid, 0x8100);
673
674         /* Receive mode register */
675         r = emac_iff2rmr(ndev);
676         if (r & EMAC_RMR_MAE)
677                 emac_hash_mc(dev);
678         out_be32(&p->rmr, r);
679
680         /* FIFOs thresholds */
681         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
682                 r = EMAC4_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
683                                tx_size / 2 / dev->fifo_entry_size);
684         else
685                 r = EMAC_TMR1((dev->mal_burst_size / dev->fifo_entry_size) + 1,
686                               tx_size / 2 / dev->fifo_entry_size);
687         out_be32(&p->tmr1, r);
688         out_be32(&p->trtr, emac_calc_trtr(dev, tx_size / 2));
689
690         /* PAUSE frame is sent when RX FIFO reaches its high-water mark,
691            there should be still enough space in FIFO to allow the our link
692            partner time to process this frame and also time to send PAUSE
693            frame itself.
694
695            Here is the worst case scenario for the RX FIFO "headroom"
696            (from "The Switch Book") (100Mbps, without preamble, inter-frame gap):
697
698            1) One maximum-length frame on TX                    1522 bytes
699            2) One PAUSE frame time                                64 bytes
700            3) PAUSE frame decode time allowance                   64 bytes
701            4) One maximum-length frame on RX                    1522 bytes
702            5) Round-trip propagation delay of the link (100Mb)    15 bytes
703            ----------
704            3187 bytes
705
706            I chose to set high-water mark to RX_FIFO_SIZE / 4 (1024 bytes)
707            low-water mark  to RX_FIFO_SIZE / 8 (512 bytes)
708          */
709         r = emac_calc_rwmr(dev, rx_size / 8 / dev->fifo_entry_size,
710                            rx_size / 4 / dev->fifo_entry_size);
711         out_be32(&p->rwmr, r);
712
713         /* Set PAUSE timer to the maximum */
714         out_be32(&p->ptr, 0xffff);
715
716         /* IRQ sources */
717         r = EMAC_ISR_OVR | EMAC_ISR_BP | EMAC_ISR_SE |
718                 EMAC_ISR_ALE | EMAC_ISR_BFCS | EMAC_ISR_PTLE | EMAC_ISR_ORE |
719                 EMAC_ISR_IRE | EMAC_ISR_TE;
720         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
721             r |= EMAC4_ISR_TXPE | EMAC4_ISR_RXPE /* | EMAC4_ISR_TXUE |
722                                                   EMAC4_ISR_RXOE | */;
723         out_be32(&p->iser,  r);
724
725         /* We need to take GPCS PHY out of isolate mode after EMAC reset */
726         if (emac_phy_gpcs(dev->phy.mode)) {
727                 if (dev->phy.gpcs_address != 0xffffffff)
728                         emac_mii_reset_gpcs(&dev->phy);
729                 else
730                         emac_mii_reset_phy(&dev->phy);
731         }
732
733         return 0;
734 }
735
736 static void emac_reinitialize(struct emac_instance *dev)
737 {
738         DBG(dev, "reinitialize" NL);
739
740         emac_netif_stop(dev);
741         if (!emac_configure(dev)) {
742                 emac_tx_enable(dev);
743                 emac_rx_enable(dev);
744         }
745         emac_netif_start(dev);
746 }
747
748 static void emac_full_tx_reset(struct emac_instance *dev)
749 {
750         DBG(dev, "full_tx_reset" NL);
751
752         emac_tx_disable(dev);
753         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
754         emac_clean_tx_ring(dev);
755         dev->tx_cnt = dev->tx_slot = dev->ack_slot = 0;
756
757         emac_configure(dev);
758
759         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
760         emac_tx_enable(dev);
761         emac_rx_enable(dev);
762 }
763
764 static void emac_reset_work(struct work_struct *work)
765 {
766         struct emac_instance *dev = container_of(work, struct emac_instance, reset_work);
767
768         DBG(dev, "reset_work" NL);
769
770         mutex_lock(&dev->link_lock);
771         if (dev->opened) {
772                 emac_netif_stop(dev);
773                 emac_full_tx_reset(dev);
774                 emac_netif_start(dev);
775         }
776         mutex_unlock(&dev->link_lock);
777 }
778
779 static void emac_tx_timeout(struct net_device *ndev, unsigned int txqueue)
780 {
781         struct emac_instance *dev = netdev_priv(ndev);
782
783         DBG(dev, "tx_timeout" NL);
784
785         schedule_work(&dev->reset_work);
786 }
787
788
789 static inline int emac_phy_done(struct emac_instance *dev, u32 stacr)
790 {
791         int done = !!(stacr & EMAC_STACR_OC);
792
793         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
794                 done = !done;
795
796         return done;
797 };
798
799 static int __emac_mdio_read(struct emac_instance *dev, u8 id, u8 reg)
800 {
801         struct emac_regs __iomem *p = dev->emacp;
802         u32 r = 0;
803         int n, err = -ETIMEDOUT;
804
805         mutex_lock(&dev->mdio_lock);
806
807         DBG2(dev, "mdio_read(%02x,%02x)" NL, id, reg);
808
809         /* Enable proper MDIO port */
810         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
811                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
812         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
813                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
814
815         /* Wait for management interface to become idle */
816         n = 20;
817         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
818                 udelay(1);
819                 if (!--n) {
820                         DBG2(dev, " -> timeout wait idle\n");
821                         goto bail;
822                 }
823         }
824
825         /* Issue read command */
826         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
827                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
828         else
829                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
830         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
831                 r |= EMAC_STACR_OC;
832         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
833                 r |= EMACX_STACR_STAC_READ;
834         else
835                 r |= EMAC_STACR_STAC_READ;
836         r |= (reg & EMAC_STACR_PRA_MASK)
837                 | ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT);
838         out_be32(&p->stacr, r);
839
840         /* Wait for read to complete */
841         n = 200;
842         while (!emac_phy_done(dev, (r = in_be32(&p->stacr)))) {
843                 udelay(1);
844                 if (!--n) {
845                         DBG2(dev, " -> timeout wait complete\n");
846                         goto bail;
847                 }
848         }
849
850         if (unlikely(r & EMAC_STACR_PHYE)) {
851                 DBG(dev, "mdio_read(%02x, %02x) failed" NL, id, reg);
852                 err = -EREMOTEIO;
853                 goto bail;
854         }
855
856         r = ((r >> EMAC_STACR_PHYD_SHIFT) & EMAC_STACR_PHYD_MASK);
857
858         DBG2(dev, "mdio_read -> %04x" NL, r);
859         err = 0;
860  bail:
861         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
862                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
863         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
864                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
865         mutex_unlock(&dev->mdio_lock);
866
867         return err == 0 ? r : err;
868 }
869
870 static void __emac_mdio_write(struct emac_instance *dev, u8 id, u8 reg,
871                               u16 val)
872 {
873         struct emac_regs __iomem *p = dev->emacp;
874         u32 r = 0;
875         int n;
876
877         mutex_lock(&dev->mdio_lock);
878
879         DBG2(dev, "mdio_write(%02x,%02x,%04x)" NL, id, reg, val);
880
881         /* Enable proper MDIO port */
882         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
883                 zmii_get_mdio(dev->zmii_dev, dev->zmii_port);
884         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
885                 rgmii_get_mdio(dev->rgmii_dev, dev->rgmii_port);
886
887         /* Wait for management interface to be idle */
888         n = 20;
889         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
890                 udelay(1);
891                 if (!--n) {
892                         DBG2(dev, " -> timeout wait idle\n");
893                         goto bail;
894                 }
895         }
896
897         /* Issue write command */
898         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
899                 r = EMAC4_STACR_BASE(dev->opb_bus_freq);
900         else
901                 r = EMAC_STACR_BASE(dev->opb_bus_freq);
902         if (emac_has_feature(dev, EMAC_FTR_STACR_OC_INVERT))
903                 r |= EMAC_STACR_OC;
904         if (emac_has_feature(dev, EMAC_FTR_HAS_NEW_STACR))
905                 r |= EMACX_STACR_STAC_WRITE;
906         else
907                 r |= EMAC_STACR_STAC_WRITE;
908         r |= (reg & EMAC_STACR_PRA_MASK) |
909                 ((id & EMAC_STACR_PCDA_MASK) << EMAC_STACR_PCDA_SHIFT) |
910                 (val << EMAC_STACR_PHYD_SHIFT);
911         out_be32(&p->stacr, r);
912
913         /* Wait for write to complete */
914         n = 200;
915         while (!emac_phy_done(dev, in_be32(&p->stacr))) {
916                 udelay(1);
917                 if (!--n) {
918                         DBG2(dev, " -> timeout wait complete\n");
919                         goto bail;
920                 }
921         }
922  bail:
923         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
924                 rgmii_put_mdio(dev->rgmii_dev, dev->rgmii_port);
925         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
926                 zmii_put_mdio(dev->zmii_dev, dev->zmii_port);
927         mutex_unlock(&dev->mdio_lock);
928 }
929
930 static int emac_mdio_read(struct net_device *ndev, int id, int reg)
931 {
932         struct emac_instance *dev = netdev_priv(ndev);
933         int res;
934
935         res = __emac_mdio_read((dev->mdio_instance &&
936                                 dev->phy.gpcs_address != id) ?
937                                 dev->mdio_instance : dev,
938                                (u8) id, (u8) reg);
939         return res;
940 }
941
942 static void emac_mdio_write(struct net_device *ndev, int id, int reg, int val)
943 {
944         struct emac_instance *dev = netdev_priv(ndev);
945
946         __emac_mdio_write((dev->mdio_instance &&
947                            dev->phy.gpcs_address != id) ?
948                            dev->mdio_instance : dev,
949                           (u8) id, (u8) reg, (u16) val);
950 }
951
952 /* Tx lock BH */
953 static void __emac_set_multicast_list(struct emac_instance *dev)
954 {
955         struct emac_regs __iomem *p = dev->emacp;
956         u32 rmr = emac_iff2rmr(dev->ndev);
957
958         DBG(dev, "__multicast %08x" NL, rmr);
959
960         /* I decided to relax register access rules here to avoid
961          * full EMAC reset.
962          *
963          * There is a real problem with EMAC4 core if we use MWSW_001 bit
964          * in MR1 register and do a full EMAC reset.
965          * One TX BD status update is delayed and, after EMAC reset, it
966          * never happens, resulting in TX hung (it'll be recovered by TX
967          * timeout handler eventually, but this is just gross).
968          * So we either have to do full TX reset or try to cheat here :)
969          *
970          * The only required change is to RX mode register, so I *think* all
971          * we need is just to stop RX channel. This seems to work on all
972          * tested SoCs.                                                --ebs
973          *
974          * If we need the full reset, we might just trigger the workqueue
975          * and do it async... a bit nasty but should work --BenH
976          */
977         dev->mcast_pending = 0;
978         emac_rx_disable(dev);
979         if (rmr & EMAC_RMR_MAE)
980                 emac_hash_mc(dev);
981         out_be32(&p->rmr, rmr);
982         emac_rx_enable(dev);
983 }
984
985 /* Tx lock BH */
986 static void emac_set_multicast_list(struct net_device *ndev)
987 {
988         struct emac_instance *dev = netdev_priv(ndev);
989
990         DBG(dev, "multicast" NL);
991
992         BUG_ON(!netif_running(dev->ndev));
993
994         if (dev->no_mcast) {
995                 dev->mcast_pending = 1;
996                 return;
997         }
998
999         mutex_lock(&dev->link_lock);
1000         __emac_set_multicast_list(dev);
1001         mutex_unlock(&dev->link_lock);
1002 }
1003
1004 static int emac_set_mac_address(struct net_device *ndev, void *sa)
1005 {
1006         struct emac_instance *dev = netdev_priv(ndev);
1007         struct sockaddr *addr = sa;
1008         struct emac_regs __iomem *p = dev->emacp;
1009
1010         if (!is_valid_ether_addr(addr->sa_data))
1011                return -EADDRNOTAVAIL;
1012
1013         mutex_lock(&dev->link_lock);
1014
1015         memcpy(ndev->dev_addr, addr->sa_data, ndev->addr_len);
1016
1017         emac_rx_disable(dev);
1018         emac_tx_disable(dev);
1019         out_be32(&p->iahr, (ndev->dev_addr[0] << 8) | ndev->dev_addr[1]);
1020         out_be32(&p->ialr, (ndev->dev_addr[2] << 24) |
1021                 (ndev->dev_addr[3] << 16) | (ndev->dev_addr[4] << 8) |
1022                 ndev->dev_addr[5]);
1023         emac_tx_enable(dev);
1024         emac_rx_enable(dev);
1025
1026         mutex_unlock(&dev->link_lock);
1027
1028         return 0;
1029 }
1030
1031 static int emac_resize_rx_ring(struct emac_instance *dev, int new_mtu)
1032 {
1033         int rx_sync_size = emac_rx_sync_size(new_mtu);
1034         int rx_skb_size = emac_rx_skb_size(new_mtu);
1035         int i, ret = 0;
1036         int mr1_jumbo_bit_change = 0;
1037
1038         mutex_lock(&dev->link_lock);
1039         emac_netif_stop(dev);
1040         emac_rx_disable(dev);
1041         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
1042
1043         if (dev->rx_sg_skb) {
1044                 ++dev->estats.rx_dropped_resize;
1045                 dev_kfree_skb(dev->rx_sg_skb);
1046                 dev->rx_sg_skb = NULL;
1047         }
1048
1049         /* Make a first pass over RX ring and mark BDs ready, dropping
1050          * non-processed packets on the way. We need this as a separate pass
1051          * to simplify error recovery in the case of allocation failure later.
1052          */
1053         for (i = 0; i < NUM_RX_BUFF; ++i) {
1054                 if (dev->rx_desc[i].ctrl & MAL_RX_CTRL_FIRST)
1055                         ++dev->estats.rx_dropped_resize;
1056
1057                 dev->rx_desc[i].data_len = 0;
1058                 dev->rx_desc[i].ctrl = MAL_RX_CTRL_EMPTY |
1059                     (i == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1060         }
1061
1062         /* Reallocate RX ring only if bigger skb buffers are required */
1063         if (rx_skb_size <= dev->rx_skb_size)
1064                 goto skip;
1065
1066         /* Second pass, allocate new skbs */
1067         for (i = 0; i < NUM_RX_BUFF; ++i) {
1068                 struct sk_buff *skb;
1069
1070                 skb = netdev_alloc_skb_ip_align(dev->ndev, rx_skb_size);
1071                 if (!skb) {
1072                         ret = -ENOMEM;
1073                         goto oom;
1074                 }
1075
1076                 BUG_ON(!dev->rx_skb[i]);
1077                 dev_kfree_skb(dev->rx_skb[i]);
1078
1079                 dev->rx_desc[i].data_ptr =
1080                     dma_map_single(&dev->ofdev->dev, skb->data - NET_IP_ALIGN,
1081                                    rx_sync_size, DMA_FROM_DEVICE)
1082                                    + NET_IP_ALIGN;
1083                 dev->rx_skb[i] = skb;
1084         }
1085  skip:
1086         /* Check if we need to change "Jumbo" bit in MR1 */
1087         if (emac_has_feature(dev, EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE)) {
1088                 mr1_jumbo_bit_change = (new_mtu > ETH_DATA_LEN) ||
1089                                 (dev->ndev->mtu > ETH_DATA_LEN);
1090         } else {
1091                 mr1_jumbo_bit_change = (new_mtu > ETH_DATA_LEN) ^
1092                                 (dev->ndev->mtu > ETH_DATA_LEN);
1093         }
1094
1095         if (mr1_jumbo_bit_change) {
1096                 /* This is to prevent starting RX channel in emac_rx_enable() */
1097                 set_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1098
1099                 dev->ndev->mtu = new_mtu;
1100                 emac_full_tx_reset(dev);
1101         }
1102
1103         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(new_mtu));
1104  oom:
1105         /* Restart RX */
1106         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1107         dev->rx_slot = 0;
1108         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1109         emac_rx_enable(dev);
1110         emac_netif_start(dev);
1111         mutex_unlock(&dev->link_lock);
1112
1113         return ret;
1114 }
1115
1116 /* Process ctx, rtnl_lock semaphore */
1117 static int emac_change_mtu(struct net_device *ndev, int new_mtu)
1118 {
1119         struct emac_instance *dev = netdev_priv(ndev);
1120         int ret = 0;
1121
1122         DBG(dev, "change_mtu(%d)" NL, new_mtu);
1123
1124         if (netif_running(ndev)) {
1125                 /* Check if we really need to reinitialize RX ring */
1126                 if (emac_rx_skb_size(ndev->mtu) != emac_rx_skb_size(new_mtu))
1127                         ret = emac_resize_rx_ring(dev, new_mtu);
1128         }
1129
1130         if (!ret) {
1131                 ndev->mtu = new_mtu;
1132                 dev->rx_skb_size = emac_rx_skb_size(new_mtu);
1133                 dev->rx_sync_size = emac_rx_sync_size(new_mtu);
1134         }
1135
1136         return ret;
1137 }
1138
1139 static void emac_clean_tx_ring(struct emac_instance *dev)
1140 {
1141         int i;
1142
1143         for (i = 0; i < NUM_TX_BUFF; ++i) {
1144                 if (dev->tx_skb[i]) {
1145                         dev_kfree_skb(dev->tx_skb[i]);
1146                         dev->tx_skb[i] = NULL;
1147                         if (dev->tx_desc[i].ctrl & MAL_TX_CTRL_READY)
1148                                 ++dev->estats.tx_dropped;
1149                 }
1150                 dev->tx_desc[i].ctrl = 0;
1151                 dev->tx_desc[i].data_ptr = 0;
1152         }
1153 }
1154
1155 static void emac_clean_rx_ring(struct emac_instance *dev)
1156 {
1157         int i;
1158
1159         for (i = 0; i < NUM_RX_BUFF; ++i)
1160                 if (dev->rx_skb[i]) {
1161                         dev->rx_desc[i].ctrl = 0;
1162                         dev_kfree_skb(dev->rx_skb[i]);
1163                         dev->rx_skb[i] = NULL;
1164                         dev->rx_desc[i].data_ptr = 0;
1165                 }
1166
1167         if (dev->rx_sg_skb) {
1168                 dev_kfree_skb(dev->rx_sg_skb);
1169                 dev->rx_sg_skb = NULL;
1170         }
1171 }
1172
1173 static int
1174 __emac_prepare_rx_skb(struct sk_buff *skb, struct emac_instance *dev, int slot)
1175 {
1176         if (unlikely(!skb))
1177                 return -ENOMEM;
1178
1179         dev->rx_skb[slot] = skb;
1180         dev->rx_desc[slot].data_len = 0;
1181
1182         dev->rx_desc[slot].data_ptr =
1183             dma_map_single(&dev->ofdev->dev, skb->data - NET_IP_ALIGN,
1184                            dev->rx_sync_size, DMA_FROM_DEVICE) + NET_IP_ALIGN;
1185         wmb();
1186         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1187             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1188
1189         return 0;
1190 }
1191
1192 static int
1193 emac_alloc_rx_skb(struct emac_instance *dev, int slot)
1194 {
1195         struct sk_buff *skb;
1196
1197         skb = __netdev_alloc_skb_ip_align(dev->ndev, dev->rx_skb_size,
1198                                           GFP_KERNEL);
1199
1200         return __emac_prepare_rx_skb(skb, dev, slot);
1201 }
1202
1203 static int
1204 emac_alloc_rx_skb_napi(struct emac_instance *dev, int slot)
1205 {
1206         struct sk_buff *skb;
1207
1208         skb = napi_alloc_skb(&dev->mal->napi, dev->rx_skb_size);
1209
1210         return __emac_prepare_rx_skb(skb, dev, slot);
1211 }
1212
1213 static void emac_print_link_status(struct emac_instance *dev)
1214 {
1215         if (netif_carrier_ok(dev->ndev))
1216                 printk(KERN_INFO "%s: link is up, %d %s%s\n",
1217                        dev->ndev->name, dev->phy.speed,
1218                        dev->phy.duplex == DUPLEX_FULL ? "FDX" : "HDX",
1219                        dev->phy.pause ? ", pause enabled" :
1220                        dev->phy.asym_pause ? ", asymmetric pause enabled" : "");
1221         else
1222                 printk(KERN_INFO "%s: link is down\n", dev->ndev->name);
1223 }
1224
1225 /* Process ctx, rtnl_lock semaphore */
1226 static int emac_open(struct net_device *ndev)
1227 {
1228         struct emac_instance *dev = netdev_priv(ndev);
1229         int err, i;
1230
1231         DBG(dev, "open" NL);
1232
1233         /* Setup error IRQ handler */
1234         err = request_irq(dev->emac_irq, emac_irq, 0, "EMAC", dev);
1235         if (err) {
1236                 printk(KERN_ERR "%s: failed to request IRQ %d\n",
1237                        ndev->name, dev->emac_irq);
1238                 return err;
1239         }
1240
1241         /* Allocate RX ring */
1242         for (i = 0; i < NUM_RX_BUFF; ++i)
1243                 if (emac_alloc_rx_skb(dev, i)) {
1244                         printk(KERN_ERR "%s: failed to allocate RX ring\n",
1245                                ndev->name);
1246                         goto oom;
1247                 }
1248
1249         dev->tx_cnt = dev->tx_slot = dev->ack_slot = dev->rx_slot = 0;
1250         clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1251         dev->rx_sg_skb = NULL;
1252
1253         mutex_lock(&dev->link_lock);
1254         dev->opened = 1;
1255
1256         /* Start PHY polling now.
1257          */
1258         if (dev->phy.address >= 0) {
1259                 int link_poll_interval;
1260                 if (dev->phy.def->ops->poll_link(&dev->phy)) {
1261                         dev->phy.def->ops->read_link(&dev->phy);
1262                         emac_rx_clk_default(dev);
1263                         netif_carrier_on(dev->ndev);
1264                         link_poll_interval = PHY_POLL_LINK_ON;
1265                 } else {
1266                         emac_rx_clk_tx(dev);
1267                         netif_carrier_off(dev->ndev);
1268                         link_poll_interval = PHY_POLL_LINK_OFF;
1269                 }
1270                 dev->link_polling = 1;
1271                 wmb();
1272                 schedule_delayed_work(&dev->link_work, link_poll_interval);
1273                 emac_print_link_status(dev);
1274         } else
1275                 netif_carrier_on(dev->ndev);
1276
1277         /* Required for Pause packet support in EMAC */
1278         dev_mc_add_global(ndev, default_mcast_addr);
1279
1280         emac_configure(dev);
1281         mal_poll_add(dev->mal, &dev->commac);
1282         mal_enable_tx_channel(dev->mal, dev->mal_tx_chan);
1283         mal_set_rcbs(dev->mal, dev->mal_rx_chan, emac_rx_size(ndev->mtu));
1284         mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1285         emac_tx_enable(dev);
1286         emac_rx_enable(dev);
1287         emac_netif_start(dev);
1288
1289         mutex_unlock(&dev->link_lock);
1290
1291         return 0;
1292  oom:
1293         emac_clean_rx_ring(dev);
1294         free_irq(dev->emac_irq, dev);
1295
1296         return -ENOMEM;
1297 }
1298
1299 /* BHs disabled */
1300 #if 0
1301 static int emac_link_differs(struct emac_instance *dev)
1302 {
1303         u32 r = in_be32(&dev->emacp->mr1);
1304
1305         int duplex = r & EMAC_MR1_FDE ? DUPLEX_FULL : DUPLEX_HALF;
1306         int speed, pause, asym_pause;
1307
1308         if (r & EMAC_MR1_MF_1000)
1309                 speed = SPEED_1000;
1310         else if (r & EMAC_MR1_MF_100)
1311                 speed = SPEED_100;
1312         else
1313                 speed = SPEED_10;
1314
1315         switch (r & (EMAC_MR1_EIFC | EMAC_MR1_APP)) {
1316         case (EMAC_MR1_EIFC | EMAC_MR1_APP):
1317                 pause = 1;
1318                 asym_pause = 0;
1319                 break;
1320         case EMAC_MR1_APP:
1321                 pause = 0;
1322                 asym_pause = 1;
1323                 break;
1324         default:
1325                 pause = asym_pause = 0;
1326         }
1327         return speed != dev->phy.speed || duplex != dev->phy.duplex ||
1328             pause != dev->phy.pause || asym_pause != dev->phy.asym_pause;
1329 }
1330 #endif
1331
1332 static void emac_link_timer(struct work_struct *work)
1333 {
1334         struct emac_instance *dev =
1335                 container_of(to_delayed_work(work),
1336                              struct emac_instance, link_work);
1337         int link_poll_interval;
1338
1339         mutex_lock(&dev->link_lock);
1340         DBG2(dev, "link timer" NL);
1341
1342         if (!dev->opened)
1343                 goto bail;
1344
1345         if (dev->phy.def->ops->poll_link(&dev->phy)) {
1346                 if (!netif_carrier_ok(dev->ndev)) {
1347                         emac_rx_clk_default(dev);
1348                         /* Get new link parameters */
1349                         dev->phy.def->ops->read_link(&dev->phy);
1350
1351                         netif_carrier_on(dev->ndev);
1352                         emac_netif_stop(dev);
1353                         emac_full_tx_reset(dev);
1354                         emac_netif_start(dev);
1355                         emac_print_link_status(dev);
1356                 }
1357                 link_poll_interval = PHY_POLL_LINK_ON;
1358         } else {
1359                 if (netif_carrier_ok(dev->ndev)) {
1360                         emac_rx_clk_tx(dev);
1361                         netif_carrier_off(dev->ndev);
1362                         netif_tx_disable(dev->ndev);
1363                         emac_reinitialize(dev);
1364                         emac_print_link_status(dev);
1365                 }
1366                 link_poll_interval = PHY_POLL_LINK_OFF;
1367         }
1368         schedule_delayed_work(&dev->link_work, link_poll_interval);
1369  bail:
1370         mutex_unlock(&dev->link_lock);
1371 }
1372
1373 static void emac_force_link_update(struct emac_instance *dev)
1374 {
1375         netif_carrier_off(dev->ndev);
1376         smp_rmb();
1377         if (dev->link_polling) {
1378                 cancel_delayed_work_sync(&dev->link_work);
1379                 if (dev->link_polling)
1380                         schedule_delayed_work(&dev->link_work,  PHY_POLL_LINK_OFF);
1381         }
1382 }
1383
1384 /* Process ctx, rtnl_lock semaphore */
1385 static int emac_close(struct net_device *ndev)
1386 {
1387         struct emac_instance *dev = netdev_priv(ndev);
1388
1389         DBG(dev, "close" NL);
1390
1391         if (dev->phy.address >= 0) {
1392                 dev->link_polling = 0;
1393                 cancel_delayed_work_sync(&dev->link_work);
1394         }
1395         mutex_lock(&dev->link_lock);
1396         emac_netif_stop(dev);
1397         dev->opened = 0;
1398         mutex_unlock(&dev->link_lock);
1399
1400         emac_rx_disable(dev);
1401         emac_tx_disable(dev);
1402         mal_disable_rx_channel(dev->mal, dev->mal_rx_chan);
1403         mal_disable_tx_channel(dev->mal, dev->mal_tx_chan);
1404         mal_poll_del(dev->mal, &dev->commac);
1405
1406         emac_clean_tx_ring(dev);
1407         emac_clean_rx_ring(dev);
1408
1409         free_irq(dev->emac_irq, dev);
1410
1411         netif_carrier_off(ndev);
1412
1413         return 0;
1414 }
1415
1416 static inline u16 emac_tx_csum(struct emac_instance *dev,
1417                                struct sk_buff *skb)
1418 {
1419         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
1420                 (skb->ip_summed == CHECKSUM_PARTIAL)) {
1421                 ++dev->stats.tx_packets_csum;
1422                 return EMAC_TX_CTRL_TAH_CSUM;
1423         }
1424         return 0;
1425 }
1426
1427 static inline netdev_tx_t emac_xmit_finish(struct emac_instance *dev, int len)
1428 {
1429         struct emac_regs __iomem *p = dev->emacp;
1430         struct net_device *ndev = dev->ndev;
1431
1432         /* Send the packet out. If the if makes a significant perf
1433          * difference, then we can store the TMR0 value in "dev"
1434          * instead
1435          */
1436         if (emac_has_feature(dev, EMAC_FTR_EMAC4))
1437                 out_be32(&p->tmr0, EMAC4_TMR0_XMIT);
1438         else
1439                 out_be32(&p->tmr0, EMAC_TMR0_XMIT);
1440
1441         if (unlikely(++dev->tx_cnt == NUM_TX_BUFF)) {
1442                 netif_stop_queue(ndev);
1443                 DBG2(dev, "stopped TX queue" NL);
1444         }
1445
1446         netif_trans_update(ndev);
1447         ++dev->stats.tx_packets;
1448         dev->stats.tx_bytes += len;
1449
1450         return NETDEV_TX_OK;
1451 }
1452
1453 /* Tx lock BH */
1454 static netdev_tx_t emac_start_xmit(struct sk_buff *skb, struct net_device *ndev)
1455 {
1456         struct emac_instance *dev = netdev_priv(ndev);
1457         unsigned int len = skb->len;
1458         int slot;
1459
1460         u16 ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1461             MAL_TX_CTRL_LAST | emac_tx_csum(dev, skb);
1462
1463         slot = dev->tx_slot++;
1464         if (dev->tx_slot == NUM_TX_BUFF) {
1465                 dev->tx_slot = 0;
1466                 ctrl |= MAL_TX_CTRL_WRAP;
1467         }
1468
1469         DBG2(dev, "xmit(%u) %d" NL, len, slot);
1470
1471         dev->tx_skb[slot] = skb;
1472         dev->tx_desc[slot].data_ptr = dma_map_single(&dev->ofdev->dev,
1473                                                      skb->data, len,
1474                                                      DMA_TO_DEVICE);
1475         dev->tx_desc[slot].data_len = (u16) len;
1476         wmb();
1477         dev->tx_desc[slot].ctrl = ctrl;
1478
1479         return emac_xmit_finish(dev, len);
1480 }
1481
1482 static inline int emac_xmit_split(struct emac_instance *dev, int slot,
1483                                   u32 pd, int len, int last, u16 base_ctrl)
1484 {
1485         while (1) {
1486                 u16 ctrl = base_ctrl;
1487                 int chunk = min(len, MAL_MAX_TX_SIZE);
1488                 len -= chunk;
1489
1490                 slot = (slot + 1) % NUM_TX_BUFF;
1491
1492                 if (last && !len)
1493                         ctrl |= MAL_TX_CTRL_LAST;
1494                 if (slot == NUM_TX_BUFF - 1)
1495                         ctrl |= MAL_TX_CTRL_WRAP;
1496
1497                 dev->tx_skb[slot] = NULL;
1498                 dev->tx_desc[slot].data_ptr = pd;
1499                 dev->tx_desc[slot].data_len = (u16) chunk;
1500                 dev->tx_desc[slot].ctrl = ctrl;
1501                 ++dev->tx_cnt;
1502
1503                 if (!len)
1504                         break;
1505
1506                 pd += chunk;
1507         }
1508         return slot;
1509 }
1510
1511 /* Tx lock BH disabled (SG version for TAH equipped EMACs) */
1512 static netdev_tx_t
1513 emac_start_xmit_sg(struct sk_buff *skb, struct net_device *ndev)
1514 {
1515         struct emac_instance *dev = netdev_priv(ndev);
1516         int nr_frags = skb_shinfo(skb)->nr_frags;
1517         int len = skb->len, chunk;
1518         int slot, i;
1519         u16 ctrl;
1520         u32 pd;
1521
1522         /* This is common "fast" path */
1523         if (likely(!nr_frags && len <= MAL_MAX_TX_SIZE))
1524                 return emac_start_xmit(skb, ndev);
1525
1526         len -= skb->data_len;
1527
1528         /* Note, this is only an *estimation*, we can still run out of empty
1529          * slots because of the additional fragmentation into
1530          * MAL_MAX_TX_SIZE-sized chunks
1531          */
1532         if (unlikely(dev->tx_cnt + nr_frags + mal_tx_chunks(len) > NUM_TX_BUFF))
1533                 goto stop_queue;
1534
1535         ctrl = EMAC_TX_CTRL_GFCS | EMAC_TX_CTRL_GP | MAL_TX_CTRL_READY |
1536             emac_tx_csum(dev, skb);
1537         slot = dev->tx_slot;
1538
1539         /* skb data */
1540         dev->tx_skb[slot] = NULL;
1541         chunk = min(len, MAL_MAX_TX_SIZE);
1542         dev->tx_desc[slot].data_ptr = pd =
1543             dma_map_single(&dev->ofdev->dev, skb->data, len, DMA_TO_DEVICE);
1544         dev->tx_desc[slot].data_len = (u16) chunk;
1545         len -= chunk;
1546         if (unlikely(len))
1547                 slot = emac_xmit_split(dev, slot, pd + chunk, len, !nr_frags,
1548                                        ctrl);
1549         /* skb fragments */
1550         for (i = 0; i < nr_frags; ++i) {
1551                 skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
1552                 len = skb_frag_size(frag);
1553
1554                 if (unlikely(dev->tx_cnt + mal_tx_chunks(len) >= NUM_TX_BUFF))
1555                         goto undo_frame;
1556
1557                 pd = skb_frag_dma_map(&dev->ofdev->dev, frag, 0, len,
1558                                       DMA_TO_DEVICE);
1559
1560                 slot = emac_xmit_split(dev, slot, pd, len, i == nr_frags - 1,
1561                                        ctrl);
1562         }
1563
1564         DBG2(dev, "xmit_sg(%u) %d - %d" NL, skb->len, dev->tx_slot, slot);
1565
1566         /* Attach skb to the last slot so we don't release it too early */
1567         dev->tx_skb[slot] = skb;
1568
1569         /* Send the packet out */
1570         if (dev->tx_slot == NUM_TX_BUFF - 1)
1571                 ctrl |= MAL_TX_CTRL_WRAP;
1572         wmb();
1573         dev->tx_desc[dev->tx_slot].ctrl = ctrl;
1574         dev->tx_slot = (slot + 1) % NUM_TX_BUFF;
1575
1576         return emac_xmit_finish(dev, skb->len);
1577
1578  undo_frame:
1579         /* Well, too bad. Our previous estimation was overly optimistic.
1580          * Undo everything.
1581          */
1582         while (slot != dev->tx_slot) {
1583                 dev->tx_desc[slot].ctrl = 0;
1584                 --dev->tx_cnt;
1585                 if (--slot < 0)
1586                         slot = NUM_TX_BUFF - 1;
1587         }
1588         ++dev->estats.tx_undo;
1589
1590  stop_queue:
1591         netif_stop_queue(ndev);
1592         DBG2(dev, "stopped TX queue" NL);
1593         return NETDEV_TX_BUSY;
1594 }
1595
1596 /* Tx lock BHs */
1597 static void emac_parse_tx_error(struct emac_instance *dev, u16 ctrl)
1598 {
1599         struct emac_error_stats *st = &dev->estats;
1600
1601         DBG(dev, "BD TX error %04x" NL, ctrl);
1602
1603         ++st->tx_bd_errors;
1604         if (ctrl & EMAC_TX_ST_BFCS)
1605                 ++st->tx_bd_bad_fcs;
1606         if (ctrl & EMAC_TX_ST_LCS)
1607                 ++st->tx_bd_carrier_loss;
1608         if (ctrl & EMAC_TX_ST_ED)
1609                 ++st->tx_bd_excessive_deferral;
1610         if (ctrl & EMAC_TX_ST_EC)
1611                 ++st->tx_bd_excessive_collisions;
1612         if (ctrl & EMAC_TX_ST_LC)
1613                 ++st->tx_bd_late_collision;
1614         if (ctrl & EMAC_TX_ST_MC)
1615                 ++st->tx_bd_multple_collisions;
1616         if (ctrl & EMAC_TX_ST_SC)
1617                 ++st->tx_bd_single_collision;
1618         if (ctrl & EMAC_TX_ST_UR)
1619                 ++st->tx_bd_underrun;
1620         if (ctrl & EMAC_TX_ST_SQE)
1621                 ++st->tx_bd_sqe;
1622 }
1623
1624 static void emac_poll_tx(void *param)
1625 {
1626         struct emac_instance *dev = param;
1627         u32 bad_mask;
1628
1629         DBG2(dev, "poll_tx, %d %d" NL, dev->tx_cnt, dev->ack_slot);
1630
1631         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
1632                 bad_mask = EMAC_IS_BAD_TX_TAH;
1633         else
1634                 bad_mask = EMAC_IS_BAD_TX;
1635
1636         netif_tx_lock_bh(dev->ndev);
1637         if (dev->tx_cnt) {
1638                 u16 ctrl;
1639                 int slot = dev->ack_slot, n = 0;
1640         again:
1641                 ctrl = dev->tx_desc[slot].ctrl;
1642                 if (!(ctrl & MAL_TX_CTRL_READY)) {
1643                         struct sk_buff *skb = dev->tx_skb[slot];
1644                         ++n;
1645
1646                         if (skb) {
1647                                 dev_kfree_skb(skb);
1648                                 dev->tx_skb[slot] = NULL;
1649                         }
1650                         slot = (slot + 1) % NUM_TX_BUFF;
1651
1652                         if (unlikely(ctrl & bad_mask))
1653                                 emac_parse_tx_error(dev, ctrl);
1654
1655                         if (--dev->tx_cnt)
1656                                 goto again;
1657                 }
1658                 if (n) {
1659                         dev->ack_slot = slot;
1660                         if (netif_queue_stopped(dev->ndev) &&
1661                             dev->tx_cnt < EMAC_TX_WAKEUP_THRESH)
1662                                 netif_wake_queue(dev->ndev);
1663
1664                         DBG2(dev, "tx %d pkts" NL, n);
1665                 }
1666         }
1667         netif_tx_unlock_bh(dev->ndev);
1668 }
1669
1670 static inline void emac_recycle_rx_skb(struct emac_instance *dev, int slot,
1671                                        int len)
1672 {
1673         struct sk_buff *skb = dev->rx_skb[slot];
1674
1675         DBG2(dev, "recycle %d %d" NL, slot, len);
1676
1677         if (len)
1678                 dma_map_single(&dev->ofdev->dev, skb->data - NET_IP_ALIGN,
1679                                SKB_DATA_ALIGN(len + NET_IP_ALIGN),
1680                                DMA_FROM_DEVICE);
1681
1682         dev->rx_desc[slot].data_len = 0;
1683         wmb();
1684         dev->rx_desc[slot].ctrl = MAL_RX_CTRL_EMPTY |
1685             (slot == (NUM_RX_BUFF - 1) ? MAL_RX_CTRL_WRAP : 0);
1686 }
1687
1688 static void emac_parse_rx_error(struct emac_instance *dev, u16 ctrl)
1689 {
1690         struct emac_error_stats *st = &dev->estats;
1691
1692         DBG(dev, "BD RX error %04x" NL, ctrl);
1693
1694         ++st->rx_bd_errors;
1695         if (ctrl & EMAC_RX_ST_OE)
1696                 ++st->rx_bd_overrun;
1697         if (ctrl & EMAC_RX_ST_BP)
1698                 ++st->rx_bd_bad_packet;
1699         if (ctrl & EMAC_RX_ST_RP)
1700                 ++st->rx_bd_runt_packet;
1701         if (ctrl & EMAC_RX_ST_SE)
1702                 ++st->rx_bd_short_event;
1703         if (ctrl & EMAC_RX_ST_AE)
1704                 ++st->rx_bd_alignment_error;
1705         if (ctrl & EMAC_RX_ST_BFCS)
1706                 ++st->rx_bd_bad_fcs;
1707         if (ctrl & EMAC_RX_ST_PTL)
1708                 ++st->rx_bd_packet_too_long;
1709         if (ctrl & EMAC_RX_ST_ORE)
1710                 ++st->rx_bd_out_of_range;
1711         if (ctrl & EMAC_RX_ST_IRE)
1712                 ++st->rx_bd_in_range;
1713 }
1714
1715 static inline void emac_rx_csum(struct emac_instance *dev,
1716                                 struct sk_buff *skb, u16 ctrl)
1717 {
1718 #ifdef CONFIG_IBM_EMAC_TAH
1719         if (!ctrl && dev->tah_dev) {
1720                 skb->ip_summed = CHECKSUM_UNNECESSARY;
1721                 ++dev->stats.rx_packets_csum;
1722         }
1723 #endif
1724 }
1725
1726 static inline int emac_rx_sg_append(struct emac_instance *dev, int slot)
1727 {
1728         if (likely(dev->rx_sg_skb != NULL)) {
1729                 int len = dev->rx_desc[slot].data_len;
1730                 int tot_len = dev->rx_sg_skb->len + len;
1731
1732                 if (unlikely(tot_len + NET_IP_ALIGN > dev->rx_skb_size)) {
1733                         ++dev->estats.rx_dropped_mtu;
1734                         dev_kfree_skb(dev->rx_sg_skb);
1735                         dev->rx_sg_skb = NULL;
1736                 } else {
1737                         memcpy(skb_tail_pointer(dev->rx_sg_skb),
1738                                          dev->rx_skb[slot]->data, len);
1739                         skb_put(dev->rx_sg_skb, len);
1740                         emac_recycle_rx_skb(dev, slot, len);
1741                         return 0;
1742                 }
1743         }
1744         emac_recycle_rx_skb(dev, slot, 0);
1745         return -1;
1746 }
1747
1748 /* NAPI poll context */
1749 static int emac_poll_rx(void *param, int budget)
1750 {
1751         struct emac_instance *dev = param;
1752         int slot = dev->rx_slot, received = 0;
1753
1754         DBG2(dev, "poll_rx(%d)" NL, budget);
1755
1756  again:
1757         while (budget > 0) {
1758                 int len;
1759                 struct sk_buff *skb;
1760                 u16 ctrl = dev->rx_desc[slot].ctrl;
1761
1762                 if (ctrl & MAL_RX_CTRL_EMPTY)
1763                         break;
1764
1765                 skb = dev->rx_skb[slot];
1766                 mb();
1767                 len = dev->rx_desc[slot].data_len;
1768
1769                 if (unlikely(!MAL_IS_SINGLE_RX(ctrl)))
1770                         goto sg;
1771
1772                 ctrl &= EMAC_BAD_RX_MASK;
1773                 if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1774                         emac_parse_rx_error(dev, ctrl);
1775                         ++dev->estats.rx_dropped_error;
1776                         emac_recycle_rx_skb(dev, slot, 0);
1777                         len = 0;
1778                         goto next;
1779                 }
1780
1781                 if (len < ETH_HLEN) {
1782                         ++dev->estats.rx_dropped_stack;
1783                         emac_recycle_rx_skb(dev, slot, len);
1784                         goto next;
1785                 }
1786
1787                 if (len && len < EMAC_RX_COPY_THRESH) {
1788                         struct sk_buff *copy_skb;
1789
1790                         copy_skb = napi_alloc_skb(&dev->mal->napi, len);
1791                         if (unlikely(!copy_skb))
1792                                 goto oom;
1793
1794                         memcpy(copy_skb->data - NET_IP_ALIGN,
1795                                skb->data - NET_IP_ALIGN,
1796                                len + NET_IP_ALIGN);
1797                         emac_recycle_rx_skb(dev, slot, len);
1798                         skb = copy_skb;
1799                 } else if (unlikely(emac_alloc_rx_skb_napi(dev, slot)))
1800                         goto oom;
1801
1802                 skb_put(skb, len);
1803         push_packet:
1804                 skb->protocol = eth_type_trans(skb, dev->ndev);
1805                 emac_rx_csum(dev, skb, ctrl);
1806
1807                 if (unlikely(netif_receive_skb(skb) == NET_RX_DROP))
1808                         ++dev->estats.rx_dropped_stack;
1809         next:
1810                 ++dev->stats.rx_packets;
1811         skip:
1812                 dev->stats.rx_bytes += len;
1813                 slot = (slot + 1) % NUM_RX_BUFF;
1814                 --budget;
1815                 ++received;
1816                 continue;
1817         sg:
1818                 if (ctrl & MAL_RX_CTRL_FIRST) {
1819                         BUG_ON(dev->rx_sg_skb);
1820                         if (unlikely(emac_alloc_rx_skb_napi(dev, slot))) {
1821                                 DBG(dev, "rx OOM %d" NL, slot);
1822                                 ++dev->estats.rx_dropped_oom;
1823                                 emac_recycle_rx_skb(dev, slot, 0);
1824                         } else {
1825                                 dev->rx_sg_skb = skb;
1826                                 skb_put(skb, len);
1827                         }
1828                 } else if (!emac_rx_sg_append(dev, slot) &&
1829                            (ctrl & MAL_RX_CTRL_LAST)) {
1830
1831                         skb = dev->rx_sg_skb;
1832                         dev->rx_sg_skb = NULL;
1833
1834                         ctrl &= EMAC_BAD_RX_MASK;
1835                         if (unlikely(ctrl && ctrl != EMAC_RX_TAH_BAD_CSUM)) {
1836                                 emac_parse_rx_error(dev, ctrl);
1837                                 ++dev->estats.rx_dropped_error;
1838                                 dev_kfree_skb(skb);
1839                                 len = 0;
1840                         } else
1841                                 goto push_packet;
1842                 }
1843                 goto skip;
1844         oom:
1845                 DBG(dev, "rx OOM %d" NL, slot);
1846                 /* Drop the packet and recycle skb */
1847                 ++dev->estats.rx_dropped_oom;
1848                 emac_recycle_rx_skb(dev, slot, 0);
1849                 goto next;
1850         }
1851
1852         if (received) {
1853                 DBG2(dev, "rx %d BDs" NL, received);
1854                 dev->rx_slot = slot;
1855         }
1856
1857         if (unlikely(budget && test_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags))) {
1858                 mb();
1859                 if (!(dev->rx_desc[slot].ctrl & MAL_RX_CTRL_EMPTY)) {
1860                         DBG2(dev, "rx restart" NL);
1861                         received = 0;
1862                         goto again;
1863                 }
1864
1865                 if (dev->rx_sg_skb) {
1866                         DBG2(dev, "dropping partial rx packet" NL);
1867                         ++dev->estats.rx_dropped_error;
1868                         dev_kfree_skb(dev->rx_sg_skb);
1869                         dev->rx_sg_skb = NULL;
1870                 }
1871
1872                 clear_bit(MAL_COMMAC_RX_STOPPED, &dev->commac.flags);
1873                 mal_enable_rx_channel(dev->mal, dev->mal_rx_chan);
1874                 emac_rx_enable(dev);
1875                 dev->rx_slot = 0;
1876         }
1877         return received;
1878 }
1879
1880 /* NAPI poll context */
1881 static int emac_peek_rx(void *param)
1882 {
1883         struct emac_instance *dev = param;
1884
1885         return !(dev->rx_desc[dev->rx_slot].ctrl & MAL_RX_CTRL_EMPTY);
1886 }
1887
1888 /* NAPI poll context */
1889 static int emac_peek_rx_sg(void *param)
1890 {
1891         struct emac_instance *dev = param;
1892
1893         int slot = dev->rx_slot;
1894         while (1) {
1895                 u16 ctrl = dev->rx_desc[slot].ctrl;
1896                 if (ctrl & MAL_RX_CTRL_EMPTY)
1897                         return 0;
1898                 else if (ctrl & MAL_RX_CTRL_LAST)
1899                         return 1;
1900
1901                 slot = (slot + 1) % NUM_RX_BUFF;
1902
1903                 /* I'm just being paranoid here :) */
1904                 if (unlikely(slot == dev->rx_slot))
1905                         return 0;
1906         }
1907 }
1908
1909 /* Hard IRQ */
1910 static void emac_rxde(void *param)
1911 {
1912         struct emac_instance *dev = param;
1913
1914         ++dev->estats.rx_stopped;
1915         emac_rx_disable_async(dev);
1916 }
1917
1918 /* Hard IRQ */
1919 static irqreturn_t emac_irq(int irq, void *dev_instance)
1920 {
1921         struct emac_instance *dev = dev_instance;
1922         struct emac_regs __iomem *p = dev->emacp;
1923         struct emac_error_stats *st = &dev->estats;
1924         u32 isr;
1925
1926         spin_lock(&dev->lock);
1927
1928         isr = in_be32(&p->isr);
1929         out_be32(&p->isr, isr);
1930
1931         DBG(dev, "isr = %08x" NL, isr);
1932
1933         if (isr & EMAC4_ISR_TXPE)
1934                 ++st->tx_parity;
1935         if (isr & EMAC4_ISR_RXPE)
1936                 ++st->rx_parity;
1937         if (isr & EMAC4_ISR_TXUE)
1938                 ++st->tx_underrun;
1939         if (isr & EMAC4_ISR_RXOE)
1940                 ++st->rx_fifo_overrun;
1941         if (isr & EMAC_ISR_OVR)
1942                 ++st->rx_overrun;
1943         if (isr & EMAC_ISR_BP)
1944                 ++st->rx_bad_packet;
1945         if (isr & EMAC_ISR_RP)
1946                 ++st->rx_runt_packet;
1947         if (isr & EMAC_ISR_SE)
1948                 ++st->rx_short_event;
1949         if (isr & EMAC_ISR_ALE)
1950                 ++st->rx_alignment_error;
1951         if (isr & EMAC_ISR_BFCS)
1952                 ++st->rx_bad_fcs;
1953         if (isr & EMAC_ISR_PTLE)
1954                 ++st->rx_packet_too_long;
1955         if (isr & EMAC_ISR_ORE)
1956                 ++st->rx_out_of_range;
1957         if (isr & EMAC_ISR_IRE)
1958                 ++st->rx_in_range;
1959         if (isr & EMAC_ISR_SQE)
1960                 ++st->tx_sqe;
1961         if (isr & EMAC_ISR_TE)
1962                 ++st->tx_errors;
1963
1964         spin_unlock(&dev->lock);
1965
1966         return IRQ_HANDLED;
1967 }
1968
1969 static struct net_device_stats *emac_stats(struct net_device *ndev)
1970 {
1971         struct emac_instance *dev = netdev_priv(ndev);
1972         struct emac_stats *st = &dev->stats;
1973         struct emac_error_stats *est = &dev->estats;
1974         struct net_device_stats *nst = &ndev->stats;
1975         unsigned long flags;
1976
1977         DBG2(dev, "stats" NL);
1978
1979         /* Compute "legacy" statistics */
1980         spin_lock_irqsave(&dev->lock, flags);
1981         nst->rx_packets = (unsigned long)st->rx_packets;
1982         nst->rx_bytes = (unsigned long)st->rx_bytes;
1983         nst->tx_packets = (unsigned long)st->tx_packets;
1984         nst->tx_bytes = (unsigned long)st->tx_bytes;
1985         nst->rx_dropped = (unsigned long)(est->rx_dropped_oom +
1986                                           est->rx_dropped_error +
1987                                           est->rx_dropped_resize +
1988                                           est->rx_dropped_mtu);
1989         nst->tx_dropped = (unsigned long)est->tx_dropped;
1990
1991         nst->rx_errors = (unsigned long)est->rx_bd_errors;
1992         nst->rx_fifo_errors = (unsigned long)(est->rx_bd_overrun +
1993                                               est->rx_fifo_overrun +
1994                                               est->rx_overrun);
1995         nst->rx_frame_errors = (unsigned long)(est->rx_bd_alignment_error +
1996                                                est->rx_alignment_error);
1997         nst->rx_crc_errors = (unsigned long)(est->rx_bd_bad_fcs +
1998                                              est->rx_bad_fcs);
1999         nst->rx_length_errors = (unsigned long)(est->rx_bd_runt_packet +
2000                                                 est->rx_bd_short_event +
2001                                                 est->rx_bd_packet_too_long +
2002                                                 est->rx_bd_out_of_range +
2003                                                 est->rx_bd_in_range +
2004                                                 est->rx_runt_packet +
2005                                                 est->rx_short_event +
2006                                                 est->rx_packet_too_long +
2007                                                 est->rx_out_of_range +
2008                                                 est->rx_in_range);
2009
2010         nst->tx_errors = (unsigned long)(est->tx_bd_errors + est->tx_errors);
2011         nst->tx_fifo_errors = (unsigned long)(est->tx_bd_underrun +
2012                                               est->tx_underrun);
2013         nst->tx_carrier_errors = (unsigned long)est->tx_bd_carrier_loss;
2014         nst->collisions = (unsigned long)(est->tx_bd_excessive_deferral +
2015                                           est->tx_bd_excessive_collisions +
2016                                           est->tx_bd_late_collision +
2017                                           est->tx_bd_multple_collisions);
2018         spin_unlock_irqrestore(&dev->lock, flags);
2019         return nst;
2020 }
2021
2022 static struct mal_commac_ops emac_commac_ops = {
2023         .poll_tx = &emac_poll_tx,
2024         .poll_rx = &emac_poll_rx,
2025         .peek_rx = &emac_peek_rx,
2026         .rxde = &emac_rxde,
2027 };
2028
2029 static struct mal_commac_ops emac_commac_sg_ops = {
2030         .poll_tx = &emac_poll_tx,
2031         .poll_rx = &emac_poll_rx,
2032         .peek_rx = &emac_peek_rx_sg,
2033         .rxde = &emac_rxde,
2034 };
2035
2036 /* Ethtool support */
2037 static int emac_ethtool_get_link_ksettings(struct net_device *ndev,
2038                                            struct ethtool_link_ksettings *cmd)
2039 {
2040         struct emac_instance *dev = netdev_priv(ndev);
2041         u32 supported, advertising;
2042
2043         supported = dev->phy.features;
2044         cmd->base.port = PORT_MII;
2045         cmd->base.phy_address = dev->phy.address;
2046
2047         mutex_lock(&dev->link_lock);
2048         advertising = dev->phy.advertising;
2049         cmd->base.autoneg = dev->phy.autoneg;
2050         cmd->base.speed = dev->phy.speed;
2051         cmd->base.duplex = dev->phy.duplex;
2052         mutex_unlock(&dev->link_lock);
2053
2054         ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.supported,
2055                                                 supported);
2056         ethtool_convert_legacy_u32_to_link_mode(cmd->link_modes.advertising,
2057                                                 advertising);
2058
2059         return 0;
2060 }
2061
2062 static int
2063 emac_ethtool_set_link_ksettings(struct net_device *ndev,
2064                                 const struct ethtool_link_ksettings *cmd)
2065 {
2066         struct emac_instance *dev = netdev_priv(ndev);
2067         u32 f = dev->phy.features;
2068         u32 advertising;
2069
2070         ethtool_convert_link_mode_to_legacy_u32(&advertising,
2071                                                 cmd->link_modes.advertising);
2072
2073         DBG(dev, "set_settings(%d, %d, %d, 0x%08x)" NL,
2074             cmd->base.autoneg, cmd->base.speed, cmd->base.duplex, advertising);
2075
2076         /* Basic sanity checks */
2077         if (dev->phy.address < 0)
2078                 return -EOPNOTSUPP;
2079         if (cmd->base.autoneg != AUTONEG_ENABLE &&
2080             cmd->base.autoneg != AUTONEG_DISABLE)
2081                 return -EINVAL;
2082         if (cmd->base.autoneg == AUTONEG_ENABLE && advertising == 0)
2083                 return -EINVAL;
2084         if (cmd->base.duplex != DUPLEX_HALF && cmd->base.duplex != DUPLEX_FULL)
2085                 return -EINVAL;
2086
2087         if (cmd->base.autoneg == AUTONEG_DISABLE) {
2088                 switch (cmd->base.speed) {
2089                 case SPEED_10:
2090                         if (cmd->base.duplex == DUPLEX_HALF &&
2091                             !(f & SUPPORTED_10baseT_Half))
2092                                 return -EINVAL;
2093                         if (cmd->base.duplex == DUPLEX_FULL &&
2094                             !(f & SUPPORTED_10baseT_Full))
2095                                 return -EINVAL;
2096                         break;
2097                 case SPEED_100:
2098                         if (cmd->base.duplex == DUPLEX_HALF &&
2099                             !(f & SUPPORTED_100baseT_Half))
2100                                 return -EINVAL;
2101                         if (cmd->base.duplex == DUPLEX_FULL &&
2102                             !(f & SUPPORTED_100baseT_Full))
2103                                 return -EINVAL;
2104                         break;
2105                 case SPEED_1000:
2106                         if (cmd->base.duplex == DUPLEX_HALF &&
2107                             !(f & SUPPORTED_1000baseT_Half))
2108                                 return -EINVAL;
2109                         if (cmd->base.duplex == DUPLEX_FULL &&
2110                             !(f & SUPPORTED_1000baseT_Full))
2111                                 return -EINVAL;
2112                         break;
2113                 default:
2114                         return -EINVAL;
2115                 }
2116
2117                 mutex_lock(&dev->link_lock);
2118                 dev->phy.def->ops->setup_forced(&dev->phy, cmd->base.speed,
2119                                                 cmd->base.duplex);
2120                 mutex_unlock(&dev->link_lock);
2121
2122         } else {
2123                 if (!(f & SUPPORTED_Autoneg))
2124                         return -EINVAL;
2125
2126                 mutex_lock(&dev->link_lock);
2127                 dev->phy.def->ops->setup_aneg(&dev->phy,
2128                                               (advertising & f) |
2129                                               (dev->phy.advertising &
2130                                                (ADVERTISED_Pause |
2131                                                 ADVERTISED_Asym_Pause)));
2132                 mutex_unlock(&dev->link_lock);
2133         }
2134         emac_force_link_update(dev);
2135
2136         return 0;
2137 }
2138
2139 static void emac_ethtool_get_ringparam(struct net_device *ndev,
2140                                        struct ethtool_ringparam *rp)
2141 {
2142         rp->rx_max_pending = rp->rx_pending = NUM_RX_BUFF;
2143         rp->tx_max_pending = rp->tx_pending = NUM_TX_BUFF;
2144 }
2145
2146 static void emac_ethtool_get_pauseparam(struct net_device *ndev,
2147                                         struct ethtool_pauseparam *pp)
2148 {
2149         struct emac_instance *dev = netdev_priv(ndev);
2150
2151         mutex_lock(&dev->link_lock);
2152         if ((dev->phy.features & SUPPORTED_Autoneg) &&
2153             (dev->phy.advertising & (ADVERTISED_Pause | ADVERTISED_Asym_Pause)))
2154                 pp->autoneg = 1;
2155
2156         if (dev->phy.duplex == DUPLEX_FULL) {
2157                 if (dev->phy.pause)
2158                         pp->rx_pause = pp->tx_pause = 1;
2159                 else if (dev->phy.asym_pause)
2160                         pp->tx_pause = 1;
2161         }
2162         mutex_unlock(&dev->link_lock);
2163 }
2164
2165 static int emac_get_regs_len(struct emac_instance *dev)
2166 {
2167                 return sizeof(struct emac_ethtool_regs_subhdr) +
2168                         sizeof(struct emac_regs);
2169 }
2170
2171 static int emac_ethtool_get_regs_len(struct net_device *ndev)
2172 {
2173         struct emac_instance *dev = netdev_priv(ndev);
2174         int size;
2175
2176         size = sizeof(struct emac_ethtool_regs_hdr) +
2177                 emac_get_regs_len(dev) + mal_get_regs_len(dev->mal);
2178         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
2179                 size += zmii_get_regs_len(dev->zmii_dev);
2180         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
2181                 size += rgmii_get_regs_len(dev->rgmii_dev);
2182         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
2183                 size += tah_get_regs_len(dev->tah_dev);
2184
2185         return size;
2186 }
2187
2188 static void *emac_dump_regs(struct emac_instance *dev, void *buf)
2189 {
2190         struct emac_ethtool_regs_subhdr *hdr = buf;
2191
2192         hdr->index = dev->cell_index;
2193         if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
2194                 hdr->version = EMAC4SYNC_ETHTOOL_REGS_VER;
2195         } else if (emac_has_feature(dev, EMAC_FTR_EMAC4)) {
2196                 hdr->version = EMAC4_ETHTOOL_REGS_VER;
2197         } else {
2198                 hdr->version = EMAC_ETHTOOL_REGS_VER;
2199         }
2200         memcpy_fromio(hdr + 1, dev->emacp, sizeof(struct emac_regs));
2201         return (void *)(hdr + 1) + sizeof(struct emac_regs);
2202 }
2203
2204 static void emac_ethtool_get_regs(struct net_device *ndev,
2205                                   struct ethtool_regs *regs, void *buf)
2206 {
2207         struct emac_instance *dev = netdev_priv(ndev);
2208         struct emac_ethtool_regs_hdr *hdr = buf;
2209
2210         hdr->components = 0;
2211         buf = hdr + 1;
2212
2213         buf = mal_dump_regs(dev->mal, buf);
2214         buf = emac_dump_regs(dev, buf);
2215         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII)) {
2216                 hdr->components |= EMAC_ETHTOOL_REGS_ZMII;
2217                 buf = zmii_dump_regs(dev->zmii_dev, buf);
2218         }
2219         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII)) {
2220                 hdr->components |= EMAC_ETHTOOL_REGS_RGMII;
2221                 buf = rgmii_dump_regs(dev->rgmii_dev, buf);
2222         }
2223         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH)) {
2224                 hdr->components |= EMAC_ETHTOOL_REGS_TAH;
2225                 buf = tah_dump_regs(dev->tah_dev, buf);
2226         }
2227 }
2228
2229 static int emac_ethtool_nway_reset(struct net_device *ndev)
2230 {
2231         struct emac_instance *dev = netdev_priv(ndev);
2232         int res = 0;
2233
2234         DBG(dev, "nway_reset" NL);
2235
2236         if (dev->phy.address < 0)
2237                 return -EOPNOTSUPP;
2238
2239         mutex_lock(&dev->link_lock);
2240         if (!dev->phy.autoneg) {
2241                 res = -EINVAL;
2242                 goto out;
2243         }
2244
2245         dev->phy.def->ops->setup_aneg(&dev->phy, dev->phy.advertising);
2246  out:
2247         mutex_unlock(&dev->link_lock);
2248         emac_force_link_update(dev);
2249         return res;
2250 }
2251
2252 static int emac_ethtool_get_sset_count(struct net_device *ndev, int stringset)
2253 {
2254         if (stringset == ETH_SS_STATS)
2255                 return EMAC_ETHTOOL_STATS_COUNT;
2256         else
2257                 return -EINVAL;
2258 }
2259
2260 static void emac_ethtool_get_strings(struct net_device *ndev, u32 stringset,
2261                                      u8 * buf)
2262 {
2263         if (stringset == ETH_SS_STATS)
2264                 memcpy(buf, &emac_stats_keys, sizeof(emac_stats_keys));
2265 }
2266
2267 static void emac_ethtool_get_ethtool_stats(struct net_device *ndev,
2268                                            struct ethtool_stats *estats,
2269                                            u64 * tmp_stats)
2270 {
2271         struct emac_instance *dev = netdev_priv(ndev);
2272
2273         memcpy(tmp_stats, &dev->stats, sizeof(dev->stats));
2274         tmp_stats += sizeof(dev->stats) / sizeof(u64);
2275         memcpy(tmp_stats, &dev->estats, sizeof(dev->estats));
2276 }
2277
2278 static void emac_ethtool_get_drvinfo(struct net_device *ndev,
2279                                      struct ethtool_drvinfo *info)
2280 {
2281         struct emac_instance *dev = netdev_priv(ndev);
2282
2283         strlcpy(info->driver, "ibm_emac", sizeof(info->driver));
2284         strlcpy(info->version, DRV_VERSION, sizeof(info->version));
2285         snprintf(info->bus_info, sizeof(info->bus_info), "PPC 4xx EMAC-%d %pOF",
2286                  dev->cell_index, dev->ofdev->dev.of_node);
2287 }
2288
2289 static const struct ethtool_ops emac_ethtool_ops = {
2290         .get_drvinfo = emac_ethtool_get_drvinfo,
2291
2292         .get_regs_len = emac_ethtool_get_regs_len,
2293         .get_regs = emac_ethtool_get_regs,
2294
2295         .nway_reset = emac_ethtool_nway_reset,
2296
2297         .get_ringparam = emac_ethtool_get_ringparam,
2298         .get_pauseparam = emac_ethtool_get_pauseparam,
2299
2300         .get_strings = emac_ethtool_get_strings,
2301         .get_sset_count = emac_ethtool_get_sset_count,
2302         .get_ethtool_stats = emac_ethtool_get_ethtool_stats,
2303
2304         .get_link = ethtool_op_get_link,
2305         .get_link_ksettings = emac_ethtool_get_link_ksettings,
2306         .set_link_ksettings = emac_ethtool_set_link_ksettings,
2307 };
2308
2309 static int emac_ioctl(struct net_device *ndev, struct ifreq *rq, int cmd)
2310 {
2311         struct emac_instance *dev = netdev_priv(ndev);
2312         struct mii_ioctl_data *data = if_mii(rq);
2313
2314         DBG(dev, "ioctl %08x" NL, cmd);
2315
2316         if (dev->phy.address < 0)
2317                 return -EOPNOTSUPP;
2318
2319         switch (cmd) {
2320         case SIOCGMIIPHY:
2321                 data->phy_id = dev->phy.address;
2322                 fallthrough;
2323         case SIOCGMIIREG:
2324                 data->val_out = emac_mdio_read(ndev, dev->phy.address,
2325                                                data->reg_num);
2326                 return 0;
2327
2328         case SIOCSMIIREG:
2329                 emac_mdio_write(ndev, dev->phy.address, data->reg_num,
2330                                 data->val_in);
2331                 return 0;
2332         default:
2333                 return -EOPNOTSUPP;
2334         }
2335 }
2336
2337 struct emac_depentry {
2338         u32                     phandle;
2339         struct device_node      *node;
2340         struct platform_device  *ofdev;
2341         void                    *drvdata;
2342 };
2343
2344 #define EMAC_DEP_MAL_IDX        0
2345 #define EMAC_DEP_ZMII_IDX       1
2346 #define EMAC_DEP_RGMII_IDX      2
2347 #define EMAC_DEP_TAH_IDX        3
2348 #define EMAC_DEP_MDIO_IDX       4
2349 #define EMAC_DEP_PREV_IDX       5
2350 #define EMAC_DEP_COUNT          6
2351
2352 static int emac_check_deps(struct emac_instance *dev,
2353                            struct emac_depentry *deps)
2354 {
2355         int i, there = 0;
2356         struct device_node *np;
2357
2358         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2359                 /* no dependency on that item, allright */
2360                 if (deps[i].phandle == 0) {
2361                         there++;
2362                         continue;
2363                 }
2364                 /* special case for blist as the dependency might go away */
2365                 if (i == EMAC_DEP_PREV_IDX) {
2366                         np = *(dev->blist - 1);
2367                         if (np == NULL) {
2368                                 deps[i].phandle = 0;
2369                                 there++;
2370                                 continue;
2371                         }
2372                         if (deps[i].node == NULL)
2373                                 deps[i].node = of_node_get(np);
2374                 }
2375                 if (deps[i].node == NULL)
2376                         deps[i].node = of_find_node_by_phandle(deps[i].phandle);
2377                 if (deps[i].node == NULL)
2378                         continue;
2379                 if (deps[i].ofdev == NULL)
2380                         deps[i].ofdev = of_find_device_by_node(deps[i].node);
2381                 if (deps[i].ofdev == NULL)
2382                         continue;
2383                 if (deps[i].drvdata == NULL)
2384                         deps[i].drvdata = platform_get_drvdata(deps[i].ofdev);
2385                 if (deps[i].drvdata != NULL)
2386                         there++;
2387         }
2388         return there == EMAC_DEP_COUNT;
2389 }
2390
2391 static void emac_put_deps(struct emac_instance *dev)
2392 {
2393         of_dev_put(dev->mal_dev);
2394         of_dev_put(dev->zmii_dev);
2395         of_dev_put(dev->rgmii_dev);
2396         of_dev_put(dev->mdio_dev);
2397         of_dev_put(dev->tah_dev);
2398 }
2399
2400 static int emac_of_bus_notify(struct notifier_block *nb, unsigned long action,
2401                               void *data)
2402 {
2403         /* We are only intereted in device addition */
2404         if (action == BUS_NOTIFY_BOUND_DRIVER)
2405                 wake_up_all(&emac_probe_wait);
2406         return 0;
2407 }
2408
2409 static struct notifier_block emac_of_bus_notifier = {
2410         .notifier_call = emac_of_bus_notify
2411 };
2412
2413 static int emac_wait_deps(struct emac_instance *dev)
2414 {
2415         struct emac_depentry deps[EMAC_DEP_COUNT];
2416         int i, err;
2417
2418         memset(&deps, 0, sizeof(deps));
2419
2420         deps[EMAC_DEP_MAL_IDX].phandle = dev->mal_ph;
2421         deps[EMAC_DEP_ZMII_IDX].phandle = dev->zmii_ph;
2422         deps[EMAC_DEP_RGMII_IDX].phandle = dev->rgmii_ph;
2423         if (dev->tah_ph)
2424                 deps[EMAC_DEP_TAH_IDX].phandle = dev->tah_ph;
2425         if (dev->mdio_ph)
2426                 deps[EMAC_DEP_MDIO_IDX].phandle = dev->mdio_ph;
2427         if (dev->blist && dev->blist > emac_boot_list)
2428                 deps[EMAC_DEP_PREV_IDX].phandle = 0xffffffffu;
2429         bus_register_notifier(&platform_bus_type, &emac_of_bus_notifier);
2430         wait_event_timeout(emac_probe_wait,
2431                            emac_check_deps(dev, deps),
2432                            EMAC_PROBE_DEP_TIMEOUT);
2433         bus_unregister_notifier(&platform_bus_type, &emac_of_bus_notifier);
2434         err = emac_check_deps(dev, deps) ? 0 : -ENODEV;
2435         for (i = 0; i < EMAC_DEP_COUNT; i++) {
2436                 of_node_put(deps[i].node);
2437                 if (err)
2438                         of_dev_put(deps[i].ofdev);
2439         }
2440         if (err == 0) {
2441                 dev->mal_dev = deps[EMAC_DEP_MAL_IDX].ofdev;
2442                 dev->zmii_dev = deps[EMAC_DEP_ZMII_IDX].ofdev;
2443                 dev->rgmii_dev = deps[EMAC_DEP_RGMII_IDX].ofdev;
2444                 dev->tah_dev = deps[EMAC_DEP_TAH_IDX].ofdev;
2445                 dev->mdio_dev = deps[EMAC_DEP_MDIO_IDX].ofdev;
2446         }
2447         of_dev_put(deps[EMAC_DEP_PREV_IDX].ofdev);
2448         return err;
2449 }
2450
2451 static int emac_read_uint_prop(struct device_node *np, const char *name,
2452                                u32 *val, int fatal)
2453 {
2454         int len;
2455         const u32 *prop = of_get_property(np, name, &len);
2456         if (prop == NULL || len < sizeof(u32)) {
2457                 if (fatal)
2458                         printk(KERN_ERR "%pOF: missing %s property\n",
2459                                np, name);
2460                 return -ENODEV;
2461         }
2462         *val = *prop;
2463         return 0;
2464 }
2465
2466 static void emac_adjust_link(struct net_device *ndev)
2467 {
2468         struct emac_instance *dev = netdev_priv(ndev);
2469         struct phy_device *phy = dev->phy_dev;
2470
2471         dev->phy.autoneg = phy->autoneg;
2472         dev->phy.speed = phy->speed;
2473         dev->phy.duplex = phy->duplex;
2474         dev->phy.pause = phy->pause;
2475         dev->phy.asym_pause = phy->asym_pause;
2476         ethtool_convert_link_mode_to_legacy_u32(&dev->phy.advertising,
2477                                                 phy->advertising);
2478 }
2479
2480 static int emac_mii_bus_read(struct mii_bus *bus, int addr, int regnum)
2481 {
2482         int ret = emac_mdio_read(bus->priv, addr, regnum);
2483         /* This is a workaround for powered down ports/phys.
2484          * In the wild, this was seen on the Cisco Meraki MX60(W).
2485          * This hardware disables ports as part of the handoff
2486          * procedure. Accessing the ports will lead to errors
2487          * (-ETIMEDOUT, -EREMOTEIO) that do more harm than good.
2488          */
2489         return ret < 0 ? 0xffff : ret;
2490 }
2491
2492 static int emac_mii_bus_write(struct mii_bus *bus, int addr,
2493                               int regnum, u16 val)
2494 {
2495         emac_mdio_write(bus->priv, addr, regnum, val);
2496         return 0;
2497 }
2498
2499 static int emac_mii_bus_reset(struct mii_bus *bus)
2500 {
2501         struct emac_instance *dev = netdev_priv(bus->priv);
2502
2503         return emac_reset(dev);
2504 }
2505
2506 static int emac_mdio_phy_start_aneg(struct mii_phy *phy,
2507                                     struct phy_device *phy_dev)
2508 {
2509         phy_dev->autoneg = phy->autoneg;
2510         phy_dev->speed = phy->speed;
2511         phy_dev->duplex = phy->duplex;
2512         ethtool_convert_legacy_u32_to_link_mode(phy_dev->advertising,
2513                                                 phy->advertising);
2514         return phy_start_aneg(phy_dev);
2515 }
2516
2517 static int emac_mdio_setup_aneg(struct mii_phy *phy, u32 advertise)
2518 {
2519         struct net_device *ndev = phy->dev;
2520         struct emac_instance *dev = netdev_priv(ndev);
2521
2522         phy->autoneg = AUTONEG_ENABLE;
2523         phy->advertising = advertise;
2524         return emac_mdio_phy_start_aneg(phy, dev->phy_dev);
2525 }
2526
2527 static int emac_mdio_setup_forced(struct mii_phy *phy, int speed, int fd)
2528 {
2529         struct net_device *ndev = phy->dev;
2530         struct emac_instance *dev = netdev_priv(ndev);
2531
2532         phy->autoneg = AUTONEG_DISABLE;
2533         phy->speed = speed;
2534         phy->duplex = fd;
2535         return emac_mdio_phy_start_aneg(phy, dev->phy_dev);
2536 }
2537
2538 static int emac_mdio_poll_link(struct mii_phy *phy)
2539 {
2540         struct net_device *ndev = phy->dev;
2541         struct emac_instance *dev = netdev_priv(ndev);
2542         int res;
2543
2544         res = phy_read_status(dev->phy_dev);
2545         if (res) {
2546                 dev_err(&dev->ofdev->dev, "link update failed (%d).", res);
2547                 return ethtool_op_get_link(ndev);
2548         }
2549
2550         return dev->phy_dev->link;
2551 }
2552
2553 static int emac_mdio_read_link(struct mii_phy *phy)
2554 {
2555         struct net_device *ndev = phy->dev;
2556         struct emac_instance *dev = netdev_priv(ndev);
2557         struct phy_device *phy_dev = dev->phy_dev;
2558         int res;
2559
2560         res = phy_read_status(phy_dev);
2561         if (res)
2562                 return res;
2563
2564         phy->speed = phy_dev->speed;
2565         phy->duplex = phy_dev->duplex;
2566         phy->pause = phy_dev->pause;
2567         phy->asym_pause = phy_dev->asym_pause;
2568         return 0;
2569 }
2570
2571 static int emac_mdio_init_phy(struct mii_phy *phy)
2572 {
2573         struct net_device *ndev = phy->dev;
2574         struct emac_instance *dev = netdev_priv(ndev);
2575
2576         phy_start(dev->phy_dev);
2577         return phy_init_hw(dev->phy_dev);
2578 }
2579
2580 static const struct mii_phy_ops emac_dt_mdio_phy_ops = {
2581         .init           = emac_mdio_init_phy,
2582         .setup_aneg     = emac_mdio_setup_aneg,
2583         .setup_forced   = emac_mdio_setup_forced,
2584         .poll_link      = emac_mdio_poll_link,
2585         .read_link      = emac_mdio_read_link,
2586 };
2587
2588 static int emac_dt_mdio_probe(struct emac_instance *dev)
2589 {
2590         struct device_node *mii_np;
2591         int res;
2592
2593         mii_np = of_get_child_by_name(dev->ofdev->dev.of_node, "mdio");
2594         if (!mii_np) {
2595                 dev_err(&dev->ofdev->dev, "no mdio definition found.");
2596                 return -ENODEV;
2597         }
2598
2599         if (!of_device_is_available(mii_np)) {
2600                 res = -ENODEV;
2601                 goto put_node;
2602         }
2603
2604         dev->mii_bus = devm_mdiobus_alloc(&dev->ofdev->dev);
2605         if (!dev->mii_bus) {
2606                 res = -ENOMEM;
2607                 goto put_node;
2608         }
2609
2610         dev->mii_bus->priv = dev->ndev;
2611         dev->mii_bus->parent = dev->ndev->dev.parent;
2612         dev->mii_bus->name = "emac_mdio";
2613         dev->mii_bus->read = &emac_mii_bus_read;
2614         dev->mii_bus->write = &emac_mii_bus_write;
2615         dev->mii_bus->reset = &emac_mii_bus_reset;
2616         snprintf(dev->mii_bus->id, MII_BUS_ID_SIZE, "%s", dev->ofdev->name);
2617         res = of_mdiobus_register(dev->mii_bus, mii_np);
2618         if (res) {
2619                 dev_err(&dev->ofdev->dev, "cannot register MDIO bus %s (%d)",
2620                         dev->mii_bus->name, res);
2621         }
2622
2623  put_node:
2624         of_node_put(mii_np);
2625         return res;
2626 }
2627
2628 static int emac_dt_phy_connect(struct emac_instance *dev,
2629                                struct device_node *phy_handle)
2630 {
2631         dev->phy.def = devm_kzalloc(&dev->ofdev->dev, sizeof(*dev->phy.def),
2632                                     GFP_KERNEL);
2633         if (!dev->phy.def)
2634                 return -ENOMEM;
2635
2636         dev->phy_dev = of_phy_connect(dev->ndev, phy_handle, &emac_adjust_link,
2637                                       0, dev->phy_mode);
2638         if (!dev->phy_dev) {
2639                 dev_err(&dev->ofdev->dev, "failed to connect to PHY.\n");
2640                 return -ENODEV;
2641         }
2642
2643         dev->phy.def->phy_id = dev->phy_dev->drv->phy_id;
2644         dev->phy.def->phy_id_mask = dev->phy_dev->drv->phy_id_mask;
2645         dev->phy.def->name = dev->phy_dev->drv->name;
2646         dev->phy.def->ops = &emac_dt_mdio_phy_ops;
2647         ethtool_convert_link_mode_to_legacy_u32(&dev->phy.features,
2648                                                 dev->phy_dev->supported);
2649         dev->phy.address = dev->phy_dev->mdio.addr;
2650         dev->phy.mode = dev->phy_dev->interface;
2651         return 0;
2652 }
2653
2654 static int emac_dt_phy_probe(struct emac_instance *dev)
2655 {
2656         struct device_node *np = dev->ofdev->dev.of_node;
2657         struct device_node *phy_handle;
2658         int res = 1;
2659
2660         phy_handle = of_parse_phandle(np, "phy-handle", 0);
2661
2662         if (phy_handle) {
2663                 res = emac_dt_mdio_probe(dev);
2664                 if (!res) {
2665                         res = emac_dt_phy_connect(dev, phy_handle);
2666                         if (res)
2667                                 mdiobus_unregister(dev->mii_bus);
2668                 }
2669         }
2670
2671         of_node_put(phy_handle);
2672         return res;
2673 }
2674
2675 static int emac_init_phy(struct emac_instance *dev)
2676 {
2677         struct device_node *np = dev->ofdev->dev.of_node;
2678         struct net_device *ndev = dev->ndev;
2679         u32 phy_map, adv;
2680         int i;
2681
2682         dev->phy.dev = ndev;
2683         dev->phy.mode = dev->phy_mode;
2684
2685         /* PHY-less configuration. */
2686         if ((dev->phy_address == 0xffffffff && dev->phy_map == 0xffffffff) ||
2687             of_phy_is_fixed_link(np)) {
2688                 emac_reset(dev);
2689
2690                 /* PHY-less configuration. */
2691                 dev->phy.address = -1;
2692                 dev->phy.features = SUPPORTED_MII;
2693                 if (emac_phy_supports_gige(dev->phy_mode))
2694                         dev->phy.features |= SUPPORTED_1000baseT_Full;
2695                 else
2696                         dev->phy.features |= SUPPORTED_100baseT_Full;
2697                 dev->phy.pause = 1;
2698
2699                 if (of_phy_is_fixed_link(np)) {
2700                         int res = emac_dt_mdio_probe(dev);
2701
2702                         if (res)
2703                                 return res;
2704
2705                         res = of_phy_register_fixed_link(np);
2706                         dev->phy_dev = of_phy_find_device(np);
2707                         if (res || !dev->phy_dev) {
2708                                 mdiobus_unregister(dev->mii_bus);
2709                                 return res ? res : -EINVAL;
2710                         }
2711                         emac_adjust_link(dev->ndev);
2712                         put_device(&dev->phy_dev->mdio.dev);
2713                 }
2714                 return 0;
2715         }
2716
2717         mutex_lock(&emac_phy_map_lock);
2718         phy_map = dev->phy_map | busy_phy_map;
2719
2720         DBG(dev, "PHY maps %08x %08x" NL, dev->phy_map, busy_phy_map);
2721
2722         dev->phy.mdio_read = emac_mdio_read;
2723         dev->phy.mdio_write = emac_mdio_write;
2724
2725         /* Enable internal clock source */
2726 #ifdef CONFIG_PPC_DCR_NATIVE
2727         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2728                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2729 #endif
2730         /* PHY clock workaround */
2731         emac_rx_clk_tx(dev);
2732
2733         /* Enable internal clock source on 440GX*/
2734 #ifdef CONFIG_PPC_DCR_NATIVE
2735         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2736                 dcri_clrset(SDR0, SDR0_MFR, 0, SDR0_MFR_ECS);
2737 #endif
2738         /* Configure EMAC with defaults so we can at least use MDIO
2739          * This is needed mostly for 440GX
2740          */
2741         if (emac_phy_gpcs(dev->phy.mode)) {
2742                 /* XXX
2743                  * Make GPCS PHY address equal to EMAC index.
2744                  * We probably should take into account busy_phy_map
2745                  * and/or phy_map here.
2746                  *
2747                  * Note that the busy_phy_map is currently global
2748                  * while it should probably be per-ASIC...
2749                  */
2750                 dev->phy.gpcs_address = dev->gpcs_address;
2751                 if (dev->phy.gpcs_address == 0xffffffff)
2752                         dev->phy.address = dev->cell_index;
2753         }
2754
2755         emac_configure(dev);
2756
2757         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII)) {
2758                 int res = emac_dt_phy_probe(dev);
2759
2760                 switch (res) {
2761                 case 1:
2762                         /* No phy-handle property configured.
2763                          * Continue with the existing phy probe
2764                          * and setup code.
2765                          */
2766                         break;
2767
2768                 case 0:
2769                         mutex_unlock(&emac_phy_map_lock);
2770                         goto init_phy;
2771
2772                 default:
2773                         mutex_unlock(&emac_phy_map_lock);
2774                         dev_err(&dev->ofdev->dev, "failed to attach dt phy (%d).\n",
2775                                 res);
2776                         return res;
2777                 }
2778         }
2779
2780         if (dev->phy_address != 0xffffffff)
2781                 phy_map = ~(1 << dev->phy_address);
2782
2783         for (i = 0; i < 0x20; phy_map >>= 1, ++i)
2784                 if (!(phy_map & 1)) {
2785                         int r;
2786                         busy_phy_map |= 1 << i;
2787
2788                         /* Quick check if there is a PHY at the address */
2789                         r = emac_mdio_read(dev->ndev, i, MII_BMCR);
2790                         if (r == 0xffff || r < 0)
2791                                 continue;
2792                         if (!emac_mii_phy_probe(&dev->phy, i))
2793                                 break;
2794                 }
2795
2796         /* Enable external clock source */
2797 #ifdef CONFIG_PPC_DCR_NATIVE
2798         if (emac_has_feature(dev, EMAC_FTR_440GX_PHY_CLK_FIX))
2799                 dcri_clrset(SDR0, SDR0_MFR, SDR0_MFR_ECS, 0);
2800 #endif
2801         mutex_unlock(&emac_phy_map_lock);
2802         if (i == 0x20) {
2803                 printk(KERN_WARNING "%pOF: can't find PHY!\n", np);
2804                 return -ENXIO;
2805         }
2806
2807  init_phy:
2808         /* Init PHY */
2809         if (dev->phy.def->ops->init)
2810                 dev->phy.def->ops->init(&dev->phy);
2811
2812         /* Disable any PHY features not supported by the platform */
2813         dev->phy.def->features &= ~dev->phy_feat_exc;
2814         dev->phy.features &= ~dev->phy_feat_exc;
2815
2816         /* Setup initial link parameters */
2817         if (dev->phy.features & SUPPORTED_Autoneg) {
2818                 adv = dev->phy.features;
2819                 if (!emac_has_feature(dev, EMAC_FTR_NO_FLOW_CONTROL_40x))
2820                         adv |= ADVERTISED_Pause | ADVERTISED_Asym_Pause;
2821                 /* Restart autonegotiation */
2822                 dev->phy.def->ops->setup_aneg(&dev->phy, adv);
2823         } else {
2824                 u32 f = dev->phy.def->features;
2825                 int speed = SPEED_10, fd = DUPLEX_HALF;
2826
2827                 /* Select highest supported speed/duplex */
2828                 if (f & SUPPORTED_1000baseT_Full) {
2829                         speed = SPEED_1000;
2830                         fd = DUPLEX_FULL;
2831                 } else if (f & SUPPORTED_1000baseT_Half)
2832                         speed = SPEED_1000;
2833                 else if (f & SUPPORTED_100baseT_Full) {
2834                         speed = SPEED_100;
2835                         fd = DUPLEX_FULL;
2836                 } else if (f & SUPPORTED_100baseT_Half)
2837                         speed = SPEED_100;
2838                 else if (f & SUPPORTED_10baseT_Full)
2839                         fd = DUPLEX_FULL;
2840
2841                 /* Force link parameters */
2842                 dev->phy.def->ops->setup_forced(&dev->phy, speed, fd);
2843         }
2844         return 0;
2845 }
2846
2847 static int emac_init_config(struct emac_instance *dev)
2848 {
2849         struct device_node *np = dev->ofdev->dev.of_node;
2850         const void *p;
2851         int err;
2852
2853         /* Read config from device-tree */
2854         if (emac_read_uint_prop(np, "mal-device", &dev->mal_ph, 1))
2855                 return -ENXIO;
2856         if (emac_read_uint_prop(np, "mal-tx-channel", &dev->mal_tx_chan, 1))
2857                 return -ENXIO;
2858         if (emac_read_uint_prop(np, "mal-rx-channel", &dev->mal_rx_chan, 1))
2859                 return -ENXIO;
2860         if (emac_read_uint_prop(np, "cell-index", &dev->cell_index, 1))
2861                 return -ENXIO;
2862         if (emac_read_uint_prop(np, "max-frame-size", &dev->max_mtu, 0))
2863                 dev->max_mtu = ETH_DATA_LEN;
2864         if (emac_read_uint_prop(np, "rx-fifo-size", &dev->rx_fifo_size, 0))
2865                 dev->rx_fifo_size = 2048;
2866         if (emac_read_uint_prop(np, "tx-fifo-size", &dev->tx_fifo_size, 0))
2867                 dev->tx_fifo_size = 2048;
2868         if (emac_read_uint_prop(np, "rx-fifo-size-gige", &dev->rx_fifo_size_gige, 0))
2869                 dev->rx_fifo_size_gige = dev->rx_fifo_size;
2870         if (emac_read_uint_prop(np, "tx-fifo-size-gige", &dev->tx_fifo_size_gige, 0))
2871                 dev->tx_fifo_size_gige = dev->tx_fifo_size;
2872         if (emac_read_uint_prop(np, "phy-address", &dev->phy_address, 0))
2873                 dev->phy_address = 0xffffffff;
2874         if (emac_read_uint_prop(np, "phy-map", &dev->phy_map, 0))
2875                 dev->phy_map = 0xffffffff;
2876         if (emac_read_uint_prop(np, "gpcs-address", &dev->gpcs_address, 0))
2877                 dev->gpcs_address = 0xffffffff;
2878         if (emac_read_uint_prop(np->parent, "clock-frequency", &dev->opb_bus_freq, 1))
2879                 return -ENXIO;
2880         if (emac_read_uint_prop(np, "tah-device", &dev->tah_ph, 0))
2881                 dev->tah_ph = 0;
2882         if (emac_read_uint_prop(np, "tah-channel", &dev->tah_port, 0))
2883                 dev->tah_port = 0;
2884         if (emac_read_uint_prop(np, "mdio-device", &dev->mdio_ph, 0))
2885                 dev->mdio_ph = 0;
2886         if (emac_read_uint_prop(np, "zmii-device", &dev->zmii_ph, 0))
2887                 dev->zmii_ph = 0;
2888         if (emac_read_uint_prop(np, "zmii-channel", &dev->zmii_port, 0))
2889                 dev->zmii_port = 0xffffffff;
2890         if (emac_read_uint_prop(np, "rgmii-device", &dev->rgmii_ph, 0))
2891                 dev->rgmii_ph = 0;
2892         if (emac_read_uint_prop(np, "rgmii-channel", &dev->rgmii_port, 0))
2893                 dev->rgmii_port = 0xffffffff;
2894         if (emac_read_uint_prop(np, "fifo-entry-size", &dev->fifo_entry_size, 0))
2895                 dev->fifo_entry_size = 16;
2896         if (emac_read_uint_prop(np, "mal-burst-size", &dev->mal_burst_size, 0))
2897                 dev->mal_burst_size = 256;
2898
2899         /* PHY mode needs some decoding */
2900         err = of_get_phy_mode(np, &dev->phy_mode);
2901         if (err)
2902                 dev->phy_mode = PHY_INTERFACE_MODE_NA;
2903
2904         /* Check EMAC version */
2905         if (of_device_is_compatible(np, "ibm,emac4sync")) {
2906                 dev->features |= (EMAC_FTR_EMAC4 | EMAC_FTR_EMAC4SYNC);
2907                 if (of_device_is_compatible(np, "ibm,emac-460ex") ||
2908                     of_device_is_compatible(np, "ibm,emac-460gt"))
2909                         dev->features |= EMAC_FTR_460EX_PHY_CLK_FIX;
2910                 if (of_device_is_compatible(np, "ibm,emac-405ex") ||
2911                     of_device_is_compatible(np, "ibm,emac-405exr"))
2912                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2913                 if (of_device_is_compatible(np, "ibm,emac-apm821xx")) {
2914                         dev->features |= (EMAC_APM821XX_REQ_JUMBO_FRAME_SIZE |
2915                                           EMAC_FTR_APM821XX_NO_HALF_DUPLEX |
2916                                           EMAC_FTR_460EX_PHY_CLK_FIX);
2917                 }
2918         } else if (of_device_is_compatible(np, "ibm,emac4")) {
2919                 dev->features |= EMAC_FTR_EMAC4;
2920                 if (of_device_is_compatible(np, "ibm,emac-440gx"))
2921                         dev->features |= EMAC_FTR_440GX_PHY_CLK_FIX;
2922         } else {
2923                 if (of_device_is_compatible(np, "ibm,emac-440ep") ||
2924                     of_device_is_compatible(np, "ibm,emac-440gr"))
2925                         dev->features |= EMAC_FTR_440EP_PHY_CLK_FIX;
2926                 if (of_device_is_compatible(np, "ibm,emac-405ez")) {
2927 #ifdef CONFIG_IBM_EMAC_NO_FLOW_CTRL
2928                         dev->features |= EMAC_FTR_NO_FLOW_CONTROL_40x;
2929 #else
2930                         printk(KERN_ERR "%pOF: Flow control not disabled!\n",
2931                                         np);
2932                         return -ENXIO;
2933 #endif
2934                 }
2935
2936         }
2937
2938         /* Fixup some feature bits based on the device tree */
2939         if (of_get_property(np, "has-inverted-stacr-oc", NULL))
2940                 dev->features |= EMAC_FTR_STACR_OC_INVERT;
2941         if (of_get_property(np, "has-new-stacr-staopc", NULL))
2942                 dev->features |= EMAC_FTR_HAS_NEW_STACR;
2943
2944         /* CAB lacks the appropriate properties */
2945         if (of_device_is_compatible(np, "ibm,emac-axon"))
2946                 dev->features |= EMAC_FTR_HAS_NEW_STACR |
2947                         EMAC_FTR_STACR_OC_INVERT;
2948
2949         /* Enable TAH/ZMII/RGMII features as found */
2950         if (dev->tah_ph != 0) {
2951 #ifdef CONFIG_IBM_EMAC_TAH
2952                 dev->features |= EMAC_FTR_HAS_TAH;
2953 #else
2954                 printk(KERN_ERR "%pOF: TAH support not enabled !\n", np);
2955                 return -ENXIO;
2956 #endif
2957         }
2958
2959         if (dev->zmii_ph != 0) {
2960 #ifdef CONFIG_IBM_EMAC_ZMII
2961                 dev->features |= EMAC_FTR_HAS_ZMII;
2962 #else
2963                 printk(KERN_ERR "%pOF: ZMII support not enabled !\n", np);
2964                 return -ENXIO;
2965 #endif
2966         }
2967
2968         if (dev->rgmii_ph != 0) {
2969 #ifdef CONFIG_IBM_EMAC_RGMII
2970                 dev->features |= EMAC_FTR_HAS_RGMII;
2971 #else
2972                 printk(KERN_ERR "%pOF: RGMII support not enabled !\n", np);
2973                 return -ENXIO;
2974 #endif
2975         }
2976
2977         /* Read MAC-address */
2978         p = of_get_property(np, "local-mac-address", NULL);
2979         if (p == NULL) {
2980                 printk(KERN_ERR "%pOF: Can't find local-mac-address property\n",
2981                        np);
2982                 return -ENXIO;
2983         }
2984         memcpy(dev->ndev->dev_addr, p, ETH_ALEN);
2985
2986         /* IAHT and GAHT filter parameterization */
2987         if (emac_has_feature(dev, EMAC_FTR_EMAC4SYNC)) {
2988                 dev->xaht_slots_shift = EMAC4SYNC_XAHT_SLOTS_SHIFT;
2989                 dev->xaht_width_shift = EMAC4SYNC_XAHT_WIDTH_SHIFT;
2990         } else {
2991                 dev->xaht_slots_shift = EMAC4_XAHT_SLOTS_SHIFT;
2992                 dev->xaht_width_shift = EMAC4_XAHT_WIDTH_SHIFT;
2993         }
2994
2995         /* This should never happen */
2996         if (WARN_ON(EMAC_XAHT_REGS(dev) > EMAC_XAHT_MAX_REGS))
2997                 return -ENXIO;
2998
2999         DBG(dev, "features     : 0x%08x / 0x%08x\n", dev->features, EMAC_FTRS_POSSIBLE);
3000         DBG(dev, "tx_fifo_size : %d (%d gige)\n", dev->tx_fifo_size, dev->tx_fifo_size_gige);
3001         DBG(dev, "rx_fifo_size : %d (%d gige)\n", dev->rx_fifo_size, dev->rx_fifo_size_gige);
3002         DBG(dev, "max_mtu      : %d\n", dev->max_mtu);
3003         DBG(dev, "OPB freq     : %d\n", dev->opb_bus_freq);
3004
3005         return 0;
3006 }
3007
3008 static const struct net_device_ops emac_netdev_ops = {
3009         .ndo_open               = emac_open,
3010         .ndo_stop               = emac_close,
3011         .ndo_get_stats          = emac_stats,
3012         .ndo_set_rx_mode        = emac_set_multicast_list,
3013         .ndo_do_ioctl           = emac_ioctl,
3014         .ndo_tx_timeout         = emac_tx_timeout,
3015         .ndo_validate_addr      = eth_validate_addr,
3016         .ndo_set_mac_address    = emac_set_mac_address,
3017         .ndo_start_xmit         = emac_start_xmit,
3018 };
3019
3020 static const struct net_device_ops emac_gige_netdev_ops = {
3021         .ndo_open               = emac_open,
3022         .ndo_stop               = emac_close,
3023         .ndo_get_stats          = emac_stats,
3024         .ndo_set_rx_mode        = emac_set_multicast_list,
3025         .ndo_do_ioctl           = emac_ioctl,
3026         .ndo_tx_timeout         = emac_tx_timeout,
3027         .ndo_validate_addr      = eth_validate_addr,
3028         .ndo_set_mac_address    = emac_set_mac_address,
3029         .ndo_start_xmit         = emac_start_xmit_sg,
3030         .ndo_change_mtu         = emac_change_mtu,
3031 };
3032
3033 static int emac_probe(struct platform_device *ofdev)
3034 {
3035         struct net_device *ndev;
3036         struct emac_instance *dev;
3037         struct device_node *np = ofdev->dev.of_node;
3038         struct device_node **blist = NULL;
3039         int err, i;
3040
3041         /* Skip unused/unwired EMACS.  We leave the check for an unused
3042          * property here for now, but new flat device trees should set a
3043          * status property to "disabled" instead.
3044          */
3045         if (of_get_property(np, "unused", NULL) || !of_device_is_available(np))
3046                 return -ENODEV;
3047
3048         /* Find ourselves in the bootlist if we are there */
3049         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
3050                 if (emac_boot_list[i] == np)
3051                         blist = &emac_boot_list[i];
3052
3053         /* Allocate our net_device structure */
3054         err = -ENOMEM;
3055         ndev = alloc_etherdev(sizeof(struct emac_instance));
3056         if (!ndev)
3057                 goto err_gone;
3058
3059         dev = netdev_priv(ndev);
3060         dev->ndev = ndev;
3061         dev->ofdev = ofdev;
3062         dev->blist = blist;
3063         SET_NETDEV_DEV(ndev, &ofdev->dev);
3064
3065         /* Initialize some embedded data structures */
3066         mutex_init(&dev->mdio_lock);
3067         mutex_init(&dev->link_lock);
3068         spin_lock_init(&dev->lock);
3069         INIT_WORK(&dev->reset_work, emac_reset_work);
3070
3071         /* Init various config data based on device-tree */
3072         err = emac_init_config(dev);
3073         if (err)
3074                 goto err_free;
3075
3076         /* Get interrupts. EMAC irq is mandatory, WOL irq is optional */
3077         dev->emac_irq = irq_of_parse_and_map(np, 0);
3078         dev->wol_irq = irq_of_parse_and_map(np, 1);
3079         if (!dev->emac_irq) {
3080                 printk(KERN_ERR "%pOF: Can't map main interrupt\n", np);
3081                 err = -ENODEV;
3082                 goto err_free;
3083         }
3084         ndev->irq = dev->emac_irq;
3085
3086         /* Map EMAC regs */
3087         // TODO : platform_get_resource() and devm_ioremap_resource()
3088         dev->emacp = of_iomap(np, 0);
3089         if (dev->emacp == NULL) {
3090                 printk(KERN_ERR "%pOF: Can't map device registers!\n", np);
3091                 err = -ENOMEM;
3092                 goto err_irq_unmap;
3093         }
3094
3095         /* Wait for dependent devices */
3096         err = emac_wait_deps(dev);
3097         if (err) {
3098                 printk(KERN_ERR
3099                        "%pOF: Timeout waiting for dependent devices\n", np);
3100                 /*  display more info about what's missing ? */
3101                 goto err_reg_unmap;
3102         }
3103         dev->mal = platform_get_drvdata(dev->mal_dev);
3104         if (dev->mdio_dev != NULL)
3105                 dev->mdio_instance = platform_get_drvdata(dev->mdio_dev);
3106
3107         /* Register with MAL */
3108         dev->commac.ops = &emac_commac_ops;
3109         dev->commac.dev = dev;
3110         dev->commac.tx_chan_mask = MAL_CHAN_MASK(dev->mal_tx_chan);
3111         dev->commac.rx_chan_mask = MAL_CHAN_MASK(dev->mal_rx_chan);
3112         err = mal_register_commac(dev->mal, &dev->commac);
3113         if (err) {
3114                 printk(KERN_ERR "%pOF: failed to register with mal %pOF!\n",
3115                        np, dev->mal_dev->dev.of_node);
3116                 goto err_rel_deps;
3117         }
3118         dev->rx_skb_size = emac_rx_skb_size(ndev->mtu);
3119         dev->rx_sync_size = emac_rx_sync_size(ndev->mtu);
3120
3121         /* Get pointers to BD rings */
3122         dev->tx_desc =
3123             dev->mal->bd_virt + mal_tx_bd_offset(dev->mal, dev->mal_tx_chan);
3124         dev->rx_desc =
3125             dev->mal->bd_virt + mal_rx_bd_offset(dev->mal, dev->mal_rx_chan);
3126
3127         DBG(dev, "tx_desc %p" NL, dev->tx_desc);
3128         DBG(dev, "rx_desc %p" NL, dev->rx_desc);
3129
3130         /* Clean rings */
3131         memset(dev->tx_desc, 0, NUM_TX_BUFF * sizeof(struct mal_descriptor));
3132         memset(dev->rx_desc, 0, NUM_RX_BUFF * sizeof(struct mal_descriptor));
3133         memset(dev->tx_skb, 0, NUM_TX_BUFF * sizeof(struct sk_buff *));
3134         memset(dev->rx_skb, 0, NUM_RX_BUFF * sizeof(struct sk_buff *));
3135
3136         /* Attach to ZMII, if needed */
3137         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII) &&
3138             (err = zmii_attach(dev->zmii_dev, dev->zmii_port, &dev->phy_mode)) != 0)
3139                 goto err_unreg_commac;
3140
3141         /* Attach to RGMII, if needed */
3142         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII) &&
3143             (err = rgmii_attach(dev->rgmii_dev, dev->rgmii_port, dev->phy_mode)) != 0)
3144                 goto err_detach_zmii;
3145
3146         /* Attach to TAH, if needed */
3147         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH) &&
3148             (err = tah_attach(dev->tah_dev, dev->tah_port)) != 0)
3149                 goto err_detach_rgmii;
3150
3151         /* Set some link defaults before we can find out real parameters */
3152         dev->phy.speed = SPEED_100;
3153         dev->phy.duplex = DUPLEX_FULL;
3154         dev->phy.autoneg = AUTONEG_DISABLE;
3155         dev->phy.pause = dev->phy.asym_pause = 0;
3156         dev->stop_timeout = STOP_TIMEOUT_100;
3157         INIT_DELAYED_WORK(&dev->link_work, emac_link_timer);
3158
3159         /* Some SoCs like APM821xx does not support Half Duplex mode. */
3160         if (emac_has_feature(dev, EMAC_FTR_APM821XX_NO_HALF_DUPLEX)) {
3161                 dev->phy_feat_exc = (SUPPORTED_1000baseT_Half |
3162                                      SUPPORTED_100baseT_Half |
3163                                      SUPPORTED_10baseT_Half);
3164         }
3165
3166         /* Find PHY if any */
3167         err = emac_init_phy(dev);
3168         if (err != 0)
3169                 goto err_detach_tah;
3170
3171         if (dev->tah_dev) {
3172                 ndev->hw_features = NETIF_F_IP_CSUM | NETIF_F_SG;
3173                 ndev->features |= ndev->hw_features | NETIF_F_RXCSUM;
3174         }
3175         ndev->watchdog_timeo = 5 * HZ;
3176         if (emac_phy_supports_gige(dev->phy_mode)) {
3177                 ndev->netdev_ops = &emac_gige_netdev_ops;
3178                 dev->commac.ops = &emac_commac_sg_ops;
3179         } else
3180                 ndev->netdev_ops = &emac_netdev_ops;
3181         ndev->ethtool_ops = &emac_ethtool_ops;
3182
3183         /* MTU range: 46 - 1500 or whatever is in OF */
3184         ndev->min_mtu = EMAC_MIN_MTU;
3185         ndev->max_mtu = dev->max_mtu;
3186
3187         netif_carrier_off(ndev);
3188
3189         err = register_netdev(ndev);
3190         if (err) {
3191                 printk(KERN_ERR "%pOF: failed to register net device (%d)!\n",
3192                        np, err);
3193                 goto err_detach_tah;
3194         }
3195
3196         /* Set our drvdata last as we don't want them visible until we are
3197          * fully initialized
3198          */
3199         wmb();
3200         platform_set_drvdata(ofdev, dev);
3201
3202         /* There's a new kid in town ! Let's tell everybody */
3203         wake_up_all(&emac_probe_wait);
3204
3205
3206         printk(KERN_INFO "%s: EMAC-%d %pOF, MAC %pM\n",
3207                ndev->name, dev->cell_index, np, ndev->dev_addr);
3208
3209         if (dev->phy_mode == PHY_INTERFACE_MODE_SGMII)
3210                 printk(KERN_NOTICE "%s: in SGMII mode\n", ndev->name);
3211
3212         if (dev->phy.address >= 0)
3213                 printk("%s: found %s PHY (0x%02x)\n", ndev->name,
3214                        dev->phy.def->name, dev->phy.address);
3215
3216         /* Life is good */
3217         return 0;
3218
3219         /* I have a bad feeling about this ... */
3220
3221  err_detach_tah:
3222         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
3223                 tah_detach(dev->tah_dev, dev->tah_port);
3224  err_detach_rgmii:
3225         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
3226                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
3227  err_detach_zmii:
3228         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
3229                 zmii_detach(dev->zmii_dev, dev->zmii_port);
3230  err_unreg_commac:
3231         mal_unregister_commac(dev->mal, &dev->commac);
3232  err_rel_deps:
3233         emac_put_deps(dev);
3234  err_reg_unmap:
3235         iounmap(dev->emacp);
3236  err_irq_unmap:
3237         if (dev->wol_irq)
3238                 irq_dispose_mapping(dev->wol_irq);
3239         if (dev->emac_irq)
3240                 irq_dispose_mapping(dev->emac_irq);
3241  err_free:
3242         free_netdev(ndev);
3243  err_gone:
3244         /* if we were on the bootlist, remove us as we won't show up and
3245          * wake up all waiters to notify them in case they were waiting
3246          * on us
3247          */
3248         if (blist) {
3249                 *blist = NULL;
3250                 wake_up_all(&emac_probe_wait);
3251         }
3252         return err;
3253 }
3254
3255 static int emac_remove(struct platform_device *ofdev)
3256 {
3257         struct emac_instance *dev = platform_get_drvdata(ofdev);
3258
3259         DBG(dev, "remove" NL);
3260
3261         unregister_netdev(dev->ndev);
3262
3263         cancel_work_sync(&dev->reset_work);
3264
3265         if (emac_has_feature(dev, EMAC_FTR_HAS_TAH))
3266                 tah_detach(dev->tah_dev, dev->tah_port);
3267         if (emac_has_feature(dev, EMAC_FTR_HAS_RGMII))
3268                 rgmii_detach(dev->rgmii_dev, dev->rgmii_port);
3269         if (emac_has_feature(dev, EMAC_FTR_HAS_ZMII))
3270                 zmii_detach(dev->zmii_dev, dev->zmii_port);
3271
3272         if (dev->phy_dev)
3273                 phy_disconnect(dev->phy_dev);
3274
3275         if (dev->mii_bus)
3276                 mdiobus_unregister(dev->mii_bus);
3277
3278         busy_phy_map &= ~(1 << dev->phy.address);
3279         DBG(dev, "busy_phy_map now %#x" NL, busy_phy_map);
3280
3281         mal_unregister_commac(dev->mal, &dev->commac);
3282         emac_put_deps(dev);
3283
3284         iounmap(dev->emacp);
3285
3286         if (dev->wol_irq)
3287                 irq_dispose_mapping(dev->wol_irq);
3288         if (dev->emac_irq)
3289                 irq_dispose_mapping(dev->emac_irq);
3290
3291         free_netdev(dev->ndev);
3292
3293         return 0;
3294 }
3295
3296 /* XXX Features in here should be replaced by properties... */
3297 static const struct of_device_id emac_match[] =
3298 {
3299         {
3300                 .type           = "network",
3301                 .compatible     = "ibm,emac",
3302         },
3303         {
3304                 .type           = "network",
3305                 .compatible     = "ibm,emac4",
3306         },
3307         {
3308                 .type           = "network",
3309                 .compatible     = "ibm,emac4sync",
3310         },
3311         {},
3312 };
3313 MODULE_DEVICE_TABLE(of, emac_match);
3314
3315 static struct platform_driver emac_driver = {
3316         .driver = {
3317                 .name = "emac",
3318                 .of_match_table = emac_match,
3319         },
3320         .probe = emac_probe,
3321         .remove = emac_remove,
3322 };
3323
3324 static void __init emac_make_bootlist(void)
3325 {
3326         struct device_node *np = NULL;
3327         int j, max, i = 0;
3328         int cell_indices[EMAC_BOOT_LIST_SIZE];
3329
3330         /* Collect EMACs */
3331         while((np = of_find_all_nodes(np)) != NULL) {
3332                 const u32 *idx;
3333
3334                 if (of_match_node(emac_match, np) == NULL)
3335                         continue;
3336                 if (of_get_property(np, "unused", NULL))
3337                         continue;
3338                 idx = of_get_property(np, "cell-index", NULL);
3339                 if (idx == NULL)
3340                         continue;
3341                 cell_indices[i] = *idx;
3342                 emac_boot_list[i++] = of_node_get(np);
3343                 if (i >= EMAC_BOOT_LIST_SIZE) {
3344                         of_node_put(np);
3345                         break;
3346                 }
3347         }
3348         max = i;
3349
3350         /* Bubble sort them (doh, what a creative algorithm :-) */
3351         for (i = 0; max > 1 && (i < (max - 1)); i++)
3352                 for (j = i; j < max; j++) {
3353                         if (cell_indices[i] > cell_indices[j]) {
3354                                 swap(emac_boot_list[i], emac_boot_list[j]);
3355                                 swap(cell_indices[i], cell_indices[j]);
3356                         }
3357                 }
3358 }
3359
3360 static int __init emac_init(void)
3361 {
3362         int rc;
3363
3364         printk(KERN_INFO DRV_DESC ", version " DRV_VERSION "\n");
3365
3366         /* Build EMAC boot list */
3367         emac_make_bootlist();
3368
3369         /* Init submodules */
3370         rc = mal_init();
3371         if (rc)
3372                 goto err;
3373         rc = zmii_init();
3374         if (rc)
3375                 goto err_mal;
3376         rc = rgmii_init();
3377         if (rc)
3378                 goto err_zmii;
3379         rc = tah_init();
3380         if (rc)
3381                 goto err_rgmii;
3382         rc = platform_driver_register(&emac_driver);
3383         if (rc)
3384                 goto err_tah;
3385
3386         return 0;
3387
3388  err_tah:
3389         tah_exit();
3390  err_rgmii:
3391         rgmii_exit();
3392  err_zmii:
3393         zmii_exit();
3394  err_mal:
3395         mal_exit();
3396  err:
3397         return rc;
3398 }
3399
3400 static void __exit emac_exit(void)
3401 {
3402         int i;
3403
3404         platform_driver_unregister(&emac_driver);
3405
3406         tah_exit();
3407         rgmii_exit();
3408         zmii_exit();
3409         mal_exit();
3410
3411         /* Destroy EMAC boot list */
3412         for (i = 0; i < EMAC_BOOT_LIST_SIZE; i++)
3413                 of_node_put(emac_boot_list[i]);
3414 }
3415
3416 module_init(emac_init);
3417 module_exit(emac_exit);