Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * drivers/net/gianfar_ethtool.c | |
3 | * | |
4 | * Gianfar Ethernet Driver | |
5 | * Ethtool support for Gianfar Enet | |
6 | * Based on e1000 ethtool support | |
7 | * | |
8 | * Author: Andy Fleming | |
4c8d3d99 | 9 | * Maintainer: Kumar Gala |
a12f801d | 10 | * Modifier: Sandeep Gopalpet <sandeep.kumar@freescale.com> |
1da177e4 | 11 | * |
a12f801d | 12 | * Copyright 2003-2006, 2008-2009 Freescale Semiconductor, Inc. |
1da177e4 | 13 | * |
6aa20a22 JG |
14 | * This software may be used and distributed according to |
15 | * the terms of the GNU Public License, Version 2, incorporated herein | |
1da177e4 LT |
16 | * by reference. |
17 | */ | |
18 | ||
1da177e4 | 19 | #include <linux/kernel.h> |
1da177e4 LT |
20 | #include <linux/string.h> |
21 | #include <linux/errno.h> | |
22 | #include <linux/slab.h> | |
23 | #include <linux/interrupt.h> | |
24 | #include <linux/init.h> | |
25 | #include <linux/delay.h> | |
26 | #include <linux/netdevice.h> | |
27 | #include <linux/etherdevice.h> | |
28 | #include <linux/skbuff.h> | |
29 | #include <linux/spinlock.h> | |
30 | #include <linux/mm.h> | |
31 | ||
32 | #include <asm/io.h> | |
33 | #include <asm/irq.h> | |
34 | #include <asm/uaccess.h> | |
35 | #include <linux/module.h> | |
1da177e4 LT |
36 | #include <linux/crc32.h> |
37 | #include <asm/types.h> | |
1da177e4 | 38 | #include <linux/ethtool.h> |
bb40dcbb AF |
39 | #include <linux/mii.h> |
40 | #include <linux/phy.h> | |
1da177e4 LT |
41 | |
42 | #include "gianfar.h" | |
43 | ||
0bbaf069 | 44 | extern void gfar_start(struct net_device *dev); |
a12f801d | 45 | extern int gfar_clean_rx_ring(struct gfar_priv_rx_q *rx_queue, int rx_work_limit); |
1da177e4 | 46 | |
bb40dcbb AF |
47 | #define GFAR_MAX_COAL_USECS 0xffff |
48 | #define GFAR_MAX_COAL_FRAMES 0xff | |
0bbaf069 | 49 | static void gfar_fill_stats(struct net_device *dev, struct ethtool_stats *dummy, |
1da177e4 | 50 | u64 * buf); |
0bbaf069 KG |
51 | static void gfar_gstrings(struct net_device *dev, u32 stringset, u8 * buf); |
52 | static int gfar_gcoalesce(struct net_device *dev, struct ethtool_coalesce *cvals); | |
53 | static int gfar_scoalesce(struct net_device *dev, struct ethtool_coalesce *cvals); | |
54 | static void gfar_gringparam(struct net_device *dev, struct ethtool_ringparam *rvals); | |
55 | static int gfar_sringparam(struct net_device *dev, struct ethtool_ringparam *rvals); | |
56 | static void gfar_gdrvinfo(struct net_device *dev, struct ethtool_drvinfo *drvinfo); | |
1da177e4 LT |
57 | |
58 | static char stat_gstrings[][ETH_GSTRING_LEN] = { | |
59 | "rx-dropped-by-kernel", | |
60 | "rx-large-frame-errors", | |
61 | "rx-short-frame-errors", | |
62 | "rx-non-octet-errors", | |
63 | "rx-crc-errors", | |
64 | "rx-overrun-errors", | |
65 | "rx-busy-errors", | |
66 | "rx-babbling-errors", | |
67 | "rx-truncated-frames", | |
68 | "ethernet-bus-error", | |
69 | "tx-babbling-errors", | |
70 | "tx-underrun-errors", | |
71 | "rx-skb-missing-errors", | |
72 | "tx-timeout-errors", | |
73 | "tx-rx-64-frames", | |
74 | "tx-rx-65-127-frames", | |
75 | "tx-rx-128-255-frames", | |
76 | "tx-rx-256-511-frames", | |
77 | "tx-rx-512-1023-frames", | |
78 | "tx-rx-1024-1518-frames", | |
79 | "tx-rx-1519-1522-good-vlan", | |
80 | "rx-bytes", | |
81 | "rx-packets", | |
82 | "rx-fcs-errors", | |
83 | "receive-multicast-packet", | |
84 | "receive-broadcast-packet", | |
85 | "rx-control-frame-packets", | |
86 | "rx-pause-frame-packets", | |
87 | "rx-unknown-op-code", | |
88 | "rx-alignment-error", | |
89 | "rx-frame-length-error", | |
90 | "rx-code-error", | |
91 | "rx-carrier-sense-error", | |
92 | "rx-undersize-packets", | |
93 | "rx-oversize-packets", | |
94 | "rx-fragmented-frames", | |
95 | "rx-jabber-frames", | |
96 | "rx-dropped-frames", | |
97 | "tx-byte-counter", | |
98 | "tx-packets", | |
99 | "tx-multicast-packets", | |
100 | "tx-broadcast-packets", | |
101 | "tx-pause-control-frames", | |
102 | "tx-deferral-packets", | |
103 | "tx-excessive-deferral-packets", | |
104 | "tx-single-collision-packets", | |
105 | "tx-multiple-collision-packets", | |
106 | "tx-late-collision-packets", | |
107 | "tx-excessive-collision-packets", | |
108 | "tx-total-collision", | |
109 | "reserved", | |
110 | "tx-dropped-frames", | |
111 | "tx-jabber-frames", | |
112 | "tx-fcs-errors", | |
113 | "tx-control-frames", | |
114 | "tx-oversize-frames", | |
115 | "tx-undersize-frames", | |
116 | "tx-fragmented-frames", | |
117 | }; | |
118 | ||
0bbaf069 KG |
119 | /* Fill in a buffer with the strings which correspond to the |
120 | * stats */ | |
121 | static void gfar_gstrings(struct net_device *dev, u32 stringset, u8 * buf) | |
122 | { | |
123 | struct gfar_private *priv = netdev_priv(dev); | |
7f7f5316 | 124 | |
b31a1d8b | 125 | if (priv->device_flags & FSL_GIANFAR_DEV_HAS_RMON) |
0bbaf069 KG |
126 | memcpy(buf, stat_gstrings, GFAR_STATS_LEN * ETH_GSTRING_LEN); |
127 | else | |
128 | memcpy(buf, stat_gstrings, | |
129 | GFAR_EXTRA_STATS_LEN * ETH_GSTRING_LEN); | |
130 | } | |
131 | ||
1da177e4 LT |
132 | /* Fill in an array of 64-bit statistics from various sources. |
133 | * This array will be appended to the end of the ethtool_stats | |
134 | * structure, and returned to user space | |
135 | */ | |
0bbaf069 | 136 | static void gfar_fill_stats(struct net_device *dev, struct ethtool_stats *dummy, u64 * buf) |
1da177e4 LT |
137 | { |
138 | int i; | |
139 | struct gfar_private *priv = netdev_priv(dev); | |
46ceb60c | 140 | struct gfar __iomem *regs = priv->gfargrp[0].regs; |
1da177e4 | 141 | u64 *extra = (u64 *) & priv->extra_stats; |
1da177e4 | 142 | |
b31a1d8b | 143 | if (priv->device_flags & FSL_GIANFAR_DEV_HAS_RMON) { |
f4983704 | 144 | u32 __iomem *rmon = (u32 __iomem *) ®s->rmon; |
0bbaf069 | 145 | struct gfar_stats *stats = (struct gfar_stats *) buf; |
1da177e4 | 146 | |
0bbaf069 | 147 | for (i = 0; i < GFAR_RMON_LEN; i++) |
cc8c6e37 | 148 | stats->rmon[i] = (u64) gfar_read(&rmon[i]); |
1da177e4 | 149 | |
0bbaf069 KG |
150 | for (i = 0; i < GFAR_EXTRA_STATS_LEN; i++) |
151 | stats->extra[i] = extra[i]; | |
152 | } else | |
153 | for (i = 0; i < GFAR_EXTRA_STATS_LEN; i++) | |
154 | buf[i] = extra[i]; | |
1da177e4 LT |
155 | } |
156 | ||
b9f2c044 | 157 | static int gfar_sset_count(struct net_device *dev, int sset) |
1da177e4 | 158 | { |
1da177e4 | 159 | struct gfar_private *priv = netdev_priv(dev); |
1da177e4 | 160 | |
b9f2c044 JG |
161 | switch (sset) { |
162 | case ETH_SS_STATS: | |
b31a1d8b | 163 | if (priv->device_flags & FSL_GIANFAR_DEV_HAS_RMON) |
b9f2c044 JG |
164 | return GFAR_STATS_LEN; |
165 | else | |
166 | return GFAR_EXTRA_STATS_LEN; | |
167 | default: | |
168 | return -EOPNOTSUPP; | |
169 | } | |
1da177e4 LT |
170 | } |
171 | ||
1da177e4 | 172 | /* Fills in the drvinfo structure with some basic info */ |
0bbaf069 | 173 | static void gfar_gdrvinfo(struct net_device *dev, struct |
1da177e4 LT |
174 | ethtool_drvinfo *drvinfo) |
175 | { | |
176 | strncpy(drvinfo->driver, DRV_NAME, GFAR_INFOSTR_LEN); | |
177 | strncpy(drvinfo->version, gfar_driver_version, GFAR_INFOSTR_LEN); | |
178 | strncpy(drvinfo->fw_version, "N/A", GFAR_INFOSTR_LEN); | |
179 | strncpy(drvinfo->bus_info, "N/A", GFAR_INFOSTR_LEN); | |
1da177e4 LT |
180 | drvinfo->regdump_len = 0; |
181 | drvinfo->eedump_len = 0; | |
182 | } | |
183 | ||
bb40dcbb AF |
184 | |
185 | static int gfar_ssettings(struct net_device *dev, struct ethtool_cmd *cmd) | |
186 | { | |
187 | struct gfar_private *priv = netdev_priv(dev); | |
188 | struct phy_device *phydev = priv->phydev; | |
189 | ||
190 | if (NULL == phydev) | |
191 | return -ENODEV; | |
192 | ||
193 | return phy_ethtool_sset(phydev, cmd); | |
194 | } | |
195 | ||
196 | ||
1da177e4 | 197 | /* Return the current settings in the ethtool_cmd structure */ |
0bbaf069 | 198 | static int gfar_gsettings(struct net_device *dev, struct ethtool_cmd *cmd) |
1da177e4 LT |
199 | { |
200 | struct gfar_private *priv = netdev_priv(dev); | |
bb40dcbb | 201 | struct phy_device *phydev = priv->phydev; |
a12f801d SG |
202 | struct gfar_priv_rx_q *rx_queue = NULL; |
203 | struct gfar_priv_tx_q *tx_queue = NULL; | |
bb40dcbb AF |
204 | |
205 | if (NULL == phydev) | |
206 | return -ENODEV; | |
fba4ed03 SG |
207 | tx_queue = priv->tx_queue[0]; |
208 | rx_queue = priv->rx_queue[0]; | |
6aa20a22 | 209 | |
fba4ed03 SG |
210 | /* etsec-1.7 and older versions have only one txic |
211 | * and rxic regs although they support multiple queues */ | |
a12f801d SG |
212 | cmd->maxtxpkt = get_icft_value(tx_queue->txic); |
213 | cmd->maxrxpkt = get_icft_value(rx_queue->rxic); | |
1da177e4 | 214 | |
bb40dcbb | 215 | return phy_ethtool_gset(phydev, cmd); |
1da177e4 LT |
216 | } |
217 | ||
218 | /* Return the length of the register structure */ | |
0bbaf069 | 219 | static int gfar_reglen(struct net_device *dev) |
1da177e4 LT |
220 | { |
221 | return sizeof (struct gfar); | |
222 | } | |
223 | ||
224 | /* Return a dump of the GFAR register space */ | |
0bbaf069 | 225 | static void gfar_get_regs(struct net_device *dev, struct ethtool_regs *regs, void *regbuf) |
1da177e4 LT |
226 | { |
227 | int i; | |
228 | struct gfar_private *priv = netdev_priv(dev); | |
46ceb60c | 229 | u32 __iomem *theregs = (u32 __iomem *) priv->gfargrp[0].regs; |
1da177e4 LT |
230 | u32 *buf = (u32 *) regbuf; |
231 | ||
232 | for (i = 0; i < sizeof (struct gfar) / sizeof (u32); i++) | |
cc8c6e37 | 233 | buf[i] = gfar_read(&theregs[i]); |
1da177e4 LT |
234 | } |
235 | ||
1da177e4 LT |
236 | /* Convert microseconds to ethernet clock ticks, which changes |
237 | * depending on what speed the controller is running at */ | |
238 | static unsigned int gfar_usecs2ticks(struct gfar_private *priv, unsigned int usecs) | |
239 | { | |
240 | unsigned int count; | |
241 | ||
242 | /* The timer is different, depending on the interface speed */ | |
bb40dcbb AF |
243 | switch (priv->phydev->speed) { |
244 | case SPEED_1000: | |
1da177e4 LT |
245 | count = GFAR_GBIT_TIME; |
246 | break; | |
bb40dcbb | 247 | case SPEED_100: |
1da177e4 LT |
248 | count = GFAR_100_TIME; |
249 | break; | |
bb40dcbb | 250 | case SPEED_10: |
1da177e4 LT |
251 | default: |
252 | count = GFAR_10_TIME; | |
253 | break; | |
254 | } | |
255 | ||
256 | /* Make sure we return a number greater than 0 | |
257 | * if usecs > 0 */ | |
258 | return ((usecs * 1000 + count - 1) / count); | |
259 | } | |
260 | ||
261 | /* Convert ethernet clock ticks to microseconds */ | |
262 | static unsigned int gfar_ticks2usecs(struct gfar_private *priv, unsigned int ticks) | |
263 | { | |
264 | unsigned int count; | |
265 | ||
266 | /* The timer is different, depending on the interface speed */ | |
bb40dcbb AF |
267 | switch (priv->phydev->speed) { |
268 | case SPEED_1000: | |
1da177e4 LT |
269 | count = GFAR_GBIT_TIME; |
270 | break; | |
bb40dcbb | 271 | case SPEED_100: |
1da177e4 LT |
272 | count = GFAR_100_TIME; |
273 | break; | |
bb40dcbb | 274 | case SPEED_10: |
1da177e4 LT |
275 | default: |
276 | count = GFAR_10_TIME; | |
277 | break; | |
278 | } | |
279 | ||
280 | /* Make sure we return a number greater than 0 */ | |
281 | /* if ticks is > 0 */ | |
282 | return ((ticks * count) / 1000); | |
283 | } | |
284 | ||
285 | /* Get the coalescing parameters, and put them in the cvals | |
286 | * structure. */ | |
0bbaf069 | 287 | static int gfar_gcoalesce(struct net_device *dev, struct ethtool_coalesce *cvals) |
1da177e4 LT |
288 | { |
289 | struct gfar_private *priv = netdev_priv(dev); | |
a12f801d SG |
290 | struct gfar_priv_rx_q *rx_queue = NULL; |
291 | struct gfar_priv_tx_q *tx_queue = NULL; | |
b46a8454 DH |
292 | unsigned long rxtime; |
293 | unsigned long rxcount; | |
294 | unsigned long txtime; | |
295 | unsigned long txcount; | |
6aa20a22 | 296 | |
b31a1d8b | 297 | if (!(priv->device_flags & FSL_GIANFAR_DEV_HAS_COALESCE)) |
0bbaf069 | 298 | return -EOPNOTSUPP; |
1da177e4 | 299 | |
bb40dcbb AF |
300 | if (NULL == priv->phydev) |
301 | return -ENODEV; | |
302 | ||
fba4ed03 SG |
303 | rx_queue = priv->rx_queue[0]; |
304 | tx_queue = priv->tx_queue[0]; | |
a12f801d SG |
305 | |
306 | rxtime = get_ictt_value(rx_queue->rxic); | |
307 | rxcount = get_icft_value(rx_queue->rxic); | |
308 | txtime = get_ictt_value(tx_queue->txic); | |
309 | txcount = get_icft_value(tx_queue->txic); | |
b46a8454 DH |
310 | cvals->rx_coalesce_usecs = gfar_ticks2usecs(priv, rxtime); |
311 | cvals->rx_max_coalesced_frames = rxcount; | |
1da177e4 | 312 | |
b46a8454 DH |
313 | cvals->tx_coalesce_usecs = gfar_ticks2usecs(priv, txtime); |
314 | cvals->tx_max_coalesced_frames = txcount; | |
1da177e4 LT |
315 | |
316 | cvals->use_adaptive_rx_coalesce = 0; | |
317 | cvals->use_adaptive_tx_coalesce = 0; | |
318 | ||
319 | cvals->pkt_rate_low = 0; | |
320 | cvals->rx_coalesce_usecs_low = 0; | |
321 | cvals->rx_max_coalesced_frames_low = 0; | |
322 | cvals->tx_coalesce_usecs_low = 0; | |
323 | cvals->tx_max_coalesced_frames_low = 0; | |
324 | ||
325 | /* When the packet rate is below pkt_rate_high but above | |
326 | * pkt_rate_low (both measured in packets per second) the | |
327 | * normal {rx,tx}_* coalescing parameters are used. | |
328 | */ | |
329 | ||
330 | /* When the packet rate is (measured in packets per second) | |
331 | * is above pkt_rate_high, the {rx,tx}_*_high parameters are | |
332 | * used. | |
333 | */ | |
334 | cvals->pkt_rate_high = 0; | |
335 | cvals->rx_coalesce_usecs_high = 0; | |
336 | cvals->rx_max_coalesced_frames_high = 0; | |
337 | cvals->tx_coalesce_usecs_high = 0; | |
338 | cvals->tx_max_coalesced_frames_high = 0; | |
339 | ||
340 | /* How often to do adaptive coalescing packet rate sampling, | |
341 | * measured in seconds. Must not be zero. | |
342 | */ | |
343 | cvals->rate_sample_interval = 0; | |
344 | ||
345 | return 0; | |
346 | } | |
347 | ||
348 | /* Change the coalescing values. | |
349 | * Both cvals->*_usecs and cvals->*_frames have to be > 0 | |
350 | * in order for coalescing to be active | |
351 | */ | |
0bbaf069 | 352 | static int gfar_scoalesce(struct net_device *dev, struct ethtool_coalesce *cvals) |
1da177e4 LT |
353 | { |
354 | struct gfar_private *priv = netdev_priv(dev); | |
46ceb60c | 355 | int i = 0; |
1da177e4 | 356 | |
b31a1d8b | 357 | if (!(priv->device_flags & FSL_GIANFAR_DEV_HAS_COALESCE)) |
0bbaf069 KG |
358 | return -EOPNOTSUPP; |
359 | ||
1da177e4 | 360 | /* Set up rx coalescing */ |
46ceb60c SG |
361 | /* As of now, we will enable/disable coalescing for all |
362 | * queues together in case of eTSEC2, this will be modified | |
363 | * along with the ethtool interface */ | |
1da177e4 | 364 | if ((cvals->rx_coalesce_usecs == 0) || |
46ceb60c SG |
365 | (cvals->rx_max_coalesced_frames == 0)) { |
366 | for (i = 0; i < priv->num_rx_queues; i++) | |
367 | priv->rx_queue[i]->rxcoalescing = 0; | |
368 | } else { | |
369 | for (i = 0; i < priv->num_rx_queues; i++) | |
370 | priv->rx_queue[i]->rxcoalescing = 1; | |
371 | } | |
1da177e4 | 372 | |
bb40dcbb AF |
373 | if (NULL == priv->phydev) |
374 | return -ENODEV; | |
375 | ||
376 | /* Check the bounds of the values */ | |
377 | if (cvals->rx_coalesce_usecs > GFAR_MAX_COAL_USECS) { | |
378 | pr_info("Coalescing is limited to %d microseconds\n", | |
379 | GFAR_MAX_COAL_USECS); | |
380 | return -EINVAL; | |
381 | } | |
382 | ||
383 | if (cvals->rx_max_coalesced_frames > GFAR_MAX_COAL_FRAMES) { | |
384 | pr_info("Coalescing is limited to %d frames\n", | |
385 | GFAR_MAX_COAL_FRAMES); | |
386 | return -EINVAL; | |
387 | } | |
388 | ||
46ceb60c SG |
389 | for (i = 0; i < priv->num_rx_queues; i++) { |
390 | priv->rx_queue[i]->rxic = mk_ic_value( | |
391 | cvals->rx_max_coalesced_frames, | |
392 | gfar_usecs2ticks(priv, cvals->rx_coalesce_usecs)); | |
393 | } | |
1da177e4 LT |
394 | |
395 | /* Set up tx coalescing */ | |
396 | if ((cvals->tx_coalesce_usecs == 0) || | |
46ceb60c SG |
397 | (cvals->tx_max_coalesced_frames == 0)) { |
398 | for (i = 0; i < priv->num_tx_queues; i++) | |
399 | priv->tx_queue[i]->txcoalescing = 0; | |
400 | } else { | |
401 | for (i = 0; i < priv->num_tx_queues; i++) | |
402 | priv->tx_queue[i]->txcoalescing = 1; | |
403 | } | |
1da177e4 | 404 | |
bb40dcbb AF |
405 | /* Check the bounds of the values */ |
406 | if (cvals->tx_coalesce_usecs > GFAR_MAX_COAL_USECS) { | |
407 | pr_info("Coalescing is limited to %d microseconds\n", | |
408 | GFAR_MAX_COAL_USECS); | |
409 | return -EINVAL; | |
410 | } | |
411 | ||
412 | if (cvals->tx_max_coalesced_frames > GFAR_MAX_COAL_FRAMES) { | |
413 | pr_info("Coalescing is limited to %d frames\n", | |
414 | GFAR_MAX_COAL_FRAMES); | |
415 | return -EINVAL; | |
416 | } | |
417 | ||
46ceb60c SG |
418 | for (i = 0; i < priv->num_tx_queues; i++) { |
419 | priv->tx_queue[i]->txic = mk_ic_value( | |
420 | cvals->tx_max_coalesced_frames, | |
421 | gfar_usecs2ticks(priv, cvals->tx_coalesce_usecs)); | |
422 | } | |
1da177e4 | 423 | |
46ceb60c | 424 | gfar_configure_coalescing(priv, 0xFF, 0xFF); |
1da177e4 LT |
425 | |
426 | return 0; | |
427 | } | |
428 | ||
429 | /* Fills in rvals with the current ring parameters. Currently, | |
430 | * rx, rx_mini, and rx_jumbo rings are the same size, as mini and | |
431 | * jumbo are ignored by the driver */ | |
0bbaf069 | 432 | static void gfar_gringparam(struct net_device *dev, struct ethtool_ringparam *rvals) |
1da177e4 LT |
433 | { |
434 | struct gfar_private *priv = netdev_priv(dev); | |
a12f801d SG |
435 | struct gfar_priv_tx_q *tx_queue = NULL; |
436 | struct gfar_priv_rx_q *rx_queue = NULL; | |
437 | ||
fba4ed03 SG |
438 | tx_queue = priv->tx_queue[0]; |
439 | rx_queue = priv->rx_queue[0]; | |
1da177e4 LT |
440 | |
441 | rvals->rx_max_pending = GFAR_RX_MAX_RING_SIZE; | |
442 | rvals->rx_mini_max_pending = GFAR_RX_MAX_RING_SIZE; | |
443 | rvals->rx_jumbo_max_pending = GFAR_RX_MAX_RING_SIZE; | |
444 | rvals->tx_max_pending = GFAR_TX_MAX_RING_SIZE; | |
445 | ||
446 | /* Values changeable by the user. The valid values are | |
447 | * in the range 1 to the "*_max_pending" counterpart above. | |
448 | */ | |
a12f801d SG |
449 | rvals->rx_pending = rx_queue->rx_ring_size; |
450 | rvals->rx_mini_pending = rx_queue->rx_ring_size; | |
451 | rvals->rx_jumbo_pending = rx_queue->rx_ring_size; | |
452 | rvals->tx_pending = tx_queue->tx_ring_size; | |
1da177e4 LT |
453 | } |
454 | ||
455 | /* Change the current ring parameters, stopping the controller if | |
456 | * necessary so that we don't mess things up while we're in | |
457 | * motion. We wait for the ring to be clean before reallocating | |
458 | * the rings. */ | |
0bbaf069 | 459 | static int gfar_sringparam(struct net_device *dev, struct ethtool_ringparam *rvals) |
1da177e4 | 460 | { |
1da177e4 | 461 | struct gfar_private *priv = netdev_priv(dev); |
fba4ed03 | 462 | int err = 0, i = 0; |
1da177e4 LT |
463 | |
464 | if (rvals->rx_pending > GFAR_RX_MAX_RING_SIZE) | |
465 | return -EINVAL; | |
466 | ||
467 | if (!is_power_of_2(rvals->rx_pending)) { | |
468 | printk("%s: Ring sizes must be a power of 2\n", | |
469 | dev->name); | |
470 | return -EINVAL; | |
471 | } | |
472 | ||
473 | if (rvals->tx_pending > GFAR_TX_MAX_RING_SIZE) | |
474 | return -EINVAL; | |
475 | ||
476 | if (!is_power_of_2(rvals->tx_pending)) { | |
477 | printk("%s: Ring sizes must be a power of 2\n", | |
478 | dev->name); | |
479 | return -EINVAL; | |
480 | } | |
481 | ||
a12f801d | 482 | |
0bbaf069 KG |
483 | if (dev->flags & IFF_UP) { |
484 | unsigned long flags; | |
1da177e4 | 485 | |
0bbaf069 KG |
486 | /* Halt TX and RX, and process the frames which |
487 | * have already been received */ | |
fba4ed03 SG |
488 | local_irq_save(flags); |
489 | lock_tx_qs(priv); | |
490 | lock_rx_qs(priv); | |
fef6108d | 491 | |
0bbaf069 | 492 | gfar_halt(dev); |
fef6108d | 493 | |
fba4ed03 SG |
494 | unlock_rx_qs(priv); |
495 | unlock_tx_qs(priv); | |
496 | local_irq_restore(flags); | |
1da177e4 | 497 | |
fba4ed03 SG |
498 | for (i = 0; i < priv->num_rx_queues; i++) |
499 | gfar_clean_rx_ring(priv->rx_queue[i], | |
500 | priv->rx_queue[i]->rx_ring_size); | |
12dea57b | 501 | |
0bbaf069 KG |
502 | /* Now we take down the rings to rebuild them */ |
503 | stop_gfar(dev); | |
504 | } | |
1da177e4 | 505 | |
0bbaf069 | 506 | /* Change the size */ |
fba4ed03 SG |
507 | for (i = 0; i < priv->num_rx_queues; i++) { |
508 | priv->rx_queue[i]->rx_ring_size = rvals->rx_pending; | |
509 | priv->tx_queue[i]->tx_ring_size = rvals->tx_pending; | |
510 | priv->tx_queue[i]->num_txbdfree = priv->tx_queue[i]->tx_ring_size; | |
511 | } | |
1da177e4 | 512 | |
0bbaf069 | 513 | /* Rebuild the rings with the new size */ |
12dea57b | 514 | if (dev->flags & IFF_UP) { |
0bbaf069 | 515 | err = startup_gfar(dev); |
fba4ed03 | 516 | netif_tx_wake_all_queues(dev); |
12dea57b | 517 | } |
0bbaf069 KG |
518 | return err; |
519 | } | |
1da177e4 | 520 | |
0bbaf069 KG |
521 | static int gfar_set_rx_csum(struct net_device *dev, uint32_t data) |
522 | { | |
523 | struct gfar_private *priv = netdev_priv(dev); | |
d87eb127 | 524 | unsigned long flags; |
fba4ed03 | 525 | int err = 0, i = 0; |
1da177e4 | 526 | |
b31a1d8b | 527 | if (!(priv->device_flags & FSL_GIANFAR_DEV_HAS_CSUM)) |
0bbaf069 KG |
528 | return -EOPNOTSUPP; |
529 | ||
a12f801d | 530 | |
0bbaf069 | 531 | if (dev->flags & IFF_UP) { |
0bbaf069 KG |
532 | /* Halt TX and RX, and process the frames which |
533 | * have already been received */ | |
fba4ed03 SG |
534 | local_irq_save(flags); |
535 | lock_tx_qs(priv); | |
536 | lock_rx_qs(priv); | |
fef6108d | 537 | |
0bbaf069 | 538 | gfar_halt(dev); |
fef6108d | 539 | |
fba4ed03 SG |
540 | unlock_tx_qs(priv); |
541 | unlock_rx_qs(priv); | |
542 | local_irq_save(flags); | |
0bbaf069 | 543 | |
fba4ed03 SG |
544 | for (i = 0; i < priv->num_rx_queues; i++) |
545 | gfar_clean_rx_ring(priv->rx_queue[i], | |
546 | priv->rx_queue[i]->rx_ring_size); | |
12dea57b | 547 | |
0bbaf069 | 548 | /* Now we take down the rings to rebuild them */ |
1da177e4 LT |
549 | stop_gfar(dev); |
550 | } | |
551 | ||
d87eb127 | 552 | spin_lock_irqsave(&priv->bflock, flags); |
0bbaf069 | 553 | priv->rx_csum_enable = data; |
d87eb127 | 554 | spin_unlock_irqrestore(&priv->bflock, flags); |
1da177e4 | 555 | |
12dea57b | 556 | if (dev->flags & IFF_UP) { |
1da177e4 | 557 | err = startup_gfar(dev); |
fba4ed03 | 558 | netif_tx_wake_all_queues(dev); |
12dea57b | 559 | } |
1da177e4 LT |
560 | return err; |
561 | } | |
562 | ||
0bbaf069 KG |
563 | static uint32_t gfar_get_rx_csum(struct net_device *dev) |
564 | { | |
565 | struct gfar_private *priv = netdev_priv(dev); | |
566 | ||
b31a1d8b | 567 | if (!(priv->device_flags & FSL_GIANFAR_DEV_HAS_CSUM)) |
0bbaf069 KG |
568 | return 0; |
569 | ||
570 | return priv->rx_csum_enable; | |
571 | } | |
572 | ||
573 | static int gfar_set_tx_csum(struct net_device *dev, uint32_t data) | |
574 | { | |
0bbaf069 KG |
575 | struct gfar_private *priv = netdev_priv(dev); |
576 | ||
b31a1d8b | 577 | if (!(priv->device_flags & FSL_GIANFAR_DEV_HAS_CSUM)) |
0bbaf069 KG |
578 | return -EOPNOTSUPP; |
579 | ||
12dea57b | 580 | netif_tx_lock_bh(dev); |
0bbaf069 KG |
581 | |
582 | if (data) | |
583 | dev->features |= NETIF_F_IP_CSUM; | |
584 | else | |
585 | dev->features &= ~NETIF_F_IP_CSUM; | |
586 | ||
12dea57b | 587 | netif_tx_unlock_bh(dev); |
0bbaf069 KG |
588 | |
589 | return 0; | |
590 | } | |
591 | ||
592 | static uint32_t gfar_get_tx_csum(struct net_device *dev) | |
593 | { | |
594 | struct gfar_private *priv = netdev_priv(dev); | |
595 | ||
b31a1d8b | 596 | if (!(priv->device_flags & FSL_GIANFAR_DEV_HAS_CSUM)) |
0bbaf069 KG |
597 | return 0; |
598 | ||
599 | return (dev->features & NETIF_F_IP_CSUM) != 0; | |
600 | } | |
601 | ||
602 | static uint32_t gfar_get_msglevel(struct net_device *dev) | |
6aa20a22 | 603 | { |
0bbaf069 KG |
604 | struct gfar_private *priv = netdev_priv(dev); |
605 | return priv->msg_enable; | |
6aa20a22 JG |
606 | } |
607 | ||
0bbaf069 | 608 | static void gfar_set_msglevel(struct net_device *dev, uint32_t data) |
6aa20a22 | 609 | { |
0bbaf069 KG |
610 | struct gfar_private *priv = netdev_priv(dev); |
611 | priv->msg_enable = data; | |
612 | } | |
613 | ||
d87eb127 SW |
614 | #ifdef CONFIG_PM |
615 | static void gfar_get_wol(struct net_device *dev, struct ethtool_wolinfo *wol) | |
616 | { | |
617 | struct gfar_private *priv = netdev_priv(dev); | |
618 | ||
b31a1d8b | 619 | if (priv->device_flags & FSL_GIANFAR_DEV_HAS_MAGIC_PACKET) { |
d87eb127 SW |
620 | wol->supported = WAKE_MAGIC; |
621 | wol->wolopts = priv->wol_en ? WAKE_MAGIC : 0; | |
622 | } else { | |
623 | wol->supported = wol->wolopts = 0; | |
624 | } | |
625 | } | |
626 | ||
627 | static int gfar_set_wol(struct net_device *dev, struct ethtool_wolinfo *wol) | |
628 | { | |
629 | struct gfar_private *priv = netdev_priv(dev); | |
630 | unsigned long flags; | |
631 | ||
b31a1d8b | 632 | if (!(priv->device_flags & FSL_GIANFAR_DEV_HAS_MAGIC_PACKET) && |
d87eb127 SW |
633 | wol->wolopts != 0) |
634 | return -EINVAL; | |
635 | ||
636 | if (wol->wolopts & ~WAKE_MAGIC) | |
637 | return -EINVAL; | |
638 | ||
639 | spin_lock_irqsave(&priv->bflock, flags); | |
640 | priv->wol_en = wol->wolopts & WAKE_MAGIC ? 1 : 0; | |
2884e5cc | 641 | device_set_wakeup_enable(&dev->dev, priv->wol_en); |
d87eb127 SW |
642 | spin_unlock_irqrestore(&priv->bflock, flags); |
643 | ||
644 | return 0; | |
645 | } | |
646 | #endif | |
0bbaf069 | 647 | |
7a8b3372 SG |
648 | static int gfar_ethflow_to_class(int flow_type, u64 *class) |
649 | { | |
650 | switch (flow_type) { | |
651 | case TCP_V4_FLOW: | |
652 | *class = CLASS_CODE_TCP_IPV4; | |
653 | break; | |
654 | case UDP_V4_FLOW: | |
655 | *class = CLASS_CODE_UDP_IPV4; | |
656 | break; | |
657 | case AH_V4_FLOW: | |
658 | case ESP_V4_FLOW: | |
659 | *class = CLASS_CODE_AH_ESP_IPV4; | |
660 | break; | |
661 | case SCTP_V4_FLOW: | |
662 | *class = CLASS_CODE_SCTP_IPV4; | |
663 | break; | |
664 | case TCP_V6_FLOW: | |
665 | *class = CLASS_CODE_TCP_IPV6; | |
666 | break; | |
667 | case UDP_V6_FLOW: | |
668 | *class = CLASS_CODE_UDP_IPV6; | |
669 | break; | |
670 | case AH_V6_FLOW: | |
671 | case ESP_V6_FLOW: | |
672 | *class = CLASS_CODE_AH_ESP_IPV6; | |
673 | break; | |
674 | case SCTP_V6_FLOW: | |
675 | *class = CLASS_CODE_SCTP_IPV6; | |
676 | break; | |
677 | default: | |
678 | return 0; | |
679 | } | |
680 | ||
681 | return 1; | |
682 | } | |
683 | ||
684 | static void ethflow_to_filer_rules (struct gfar_private *priv, u64 ethflow) | |
685 | { | |
686 | u32 fcr = 0x0, fpr = FPR_FILER_MASK; | |
687 | ||
688 | if (ethflow & RXH_L2DA) { | |
689 | fcr = RQFCR_PID_DAH |RQFCR_CMP_NOMATCH | | |
690 | RQFCR_HASH | RQFCR_AND | RQFCR_HASHTBL_0; | |
691 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
692 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
693 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
694 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
695 | ||
696 | fcr = RQFCR_PID_DAL | RQFCR_AND | RQFCR_CMP_NOMATCH | | |
697 | RQFCR_HASH | RQFCR_AND | RQFCR_HASHTBL_0; | |
698 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
699 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
700 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
701 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
702 | } | |
703 | ||
704 | if (ethflow & RXH_VLAN) { | |
705 | fcr = RQFCR_PID_VID | RQFCR_CMP_NOMATCH | RQFCR_HASH | | |
706 | RQFCR_AND | RQFCR_HASHTBL_0; | |
707 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
708 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
709 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
710 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
711 | } | |
712 | ||
713 | if (ethflow & RXH_IP_SRC) { | |
714 | fcr = RQFCR_PID_SIA | RQFCR_CMP_NOMATCH | RQFCR_HASH | | |
715 | RQFCR_AND | RQFCR_HASHTBL_0; | |
716 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
717 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
718 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
719 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
720 | } | |
721 | ||
722 | if (ethflow & (RXH_IP_DST)) { | |
723 | fcr = RQFCR_PID_DIA | RQFCR_CMP_NOMATCH | RQFCR_HASH | | |
724 | RQFCR_AND | RQFCR_HASHTBL_0; | |
725 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
726 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
727 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
728 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
729 | } | |
730 | ||
731 | if (ethflow & RXH_L3_PROTO) { | |
732 | fcr = RQFCR_PID_L4P | RQFCR_CMP_NOMATCH | RQFCR_HASH | | |
733 | RQFCR_AND | RQFCR_HASHTBL_0; | |
734 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
735 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
736 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
737 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
738 | } | |
739 | ||
740 | if (ethflow & RXH_L4_B_0_1) { | |
741 | fcr = RQFCR_PID_SPT | RQFCR_CMP_NOMATCH | RQFCR_HASH | | |
742 | RQFCR_AND | RQFCR_HASHTBL_0; | |
743 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
744 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
745 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
746 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
747 | } | |
748 | ||
749 | if (ethflow & RXH_L4_B_2_3) { | |
750 | fcr = RQFCR_PID_DPT | RQFCR_CMP_NOMATCH | RQFCR_HASH | | |
751 | RQFCR_AND | RQFCR_HASHTBL_0; | |
752 | ftp_rqfpr[priv->cur_filer_idx] = fpr; | |
753 | ftp_rqfcr[priv->cur_filer_idx] = fcr; | |
754 | gfar_write_filer(priv, priv->cur_filer_idx, fcr, fpr); | |
755 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
756 | } | |
757 | } | |
758 | ||
759 | static int gfar_ethflow_to_filer_table(struct gfar_private *priv, u64 ethflow, u64 class) | |
760 | { | |
761 | unsigned int last_rule_idx = priv->cur_filer_idx; | |
762 | unsigned int cmp_rqfpr; | |
763 | unsigned int local_rqfpr[MAX_FILER_IDX + 1]; | |
764 | unsigned int local_rqfcr[MAX_FILER_IDX + 1]; | |
765 | int i = 0x0, k = 0x0; | |
766 | int j = MAX_FILER_IDX, l = 0x0; | |
767 | ||
768 | switch (class) { | |
769 | case TCP_V4_FLOW: | |
770 | cmp_rqfpr = RQFPR_IPV4 |RQFPR_TCP; | |
771 | break; | |
772 | case UDP_V4_FLOW: | |
773 | cmp_rqfpr = RQFPR_IPV4 |RQFPR_UDP; | |
774 | break; | |
775 | case TCP_V6_FLOW: | |
776 | cmp_rqfpr = RQFPR_IPV6 |RQFPR_TCP; | |
777 | break; | |
778 | case UDP_V6_FLOW: | |
779 | cmp_rqfpr = RQFPR_IPV6 |RQFPR_UDP; | |
780 | break; | |
781 | case IPV4_FLOW: | |
782 | cmp_rqfpr = RQFPR_IPV4; | |
783 | case IPV6_FLOW: | |
784 | cmp_rqfpr = RQFPR_IPV6; | |
785 | break; | |
786 | default: | |
787 | printk(KERN_ERR "Right now this class is not supported\n"); | |
788 | return 0; | |
789 | } | |
790 | ||
791 | for (i = 0; i < MAX_FILER_IDX + 1; i++) { | |
792 | local_rqfpr[j] = ftp_rqfpr[i]; | |
793 | local_rqfcr[j] = ftp_rqfcr[i]; | |
794 | j--; | |
795 | if ((ftp_rqfcr[i] == (RQFCR_PID_PARSE | | |
796 | RQFCR_CLE |RQFCR_AND)) && | |
797 | (ftp_rqfpr[i] == cmp_rqfpr)) | |
798 | break; | |
799 | } | |
800 | ||
801 | if (i == MAX_FILER_IDX + 1) { | |
802 | printk(KERN_ERR "No parse rule found, "); | |
803 | printk(KERN_ERR "can't create hash rules\n"); | |
804 | return 0; | |
805 | } | |
806 | ||
807 | /* If a match was found, then it begins the starting of a cluster rule | |
808 | * if it was already programmed, we need to overwrite these rules | |
809 | */ | |
810 | for (l = i+1; l < MAX_FILER_IDX; l++) { | |
811 | if ((ftp_rqfcr[l] & RQFCR_CLE) && | |
812 | !(ftp_rqfcr[l] & RQFCR_AND)) { | |
813 | ftp_rqfcr[l] = RQFCR_CLE | RQFCR_CMP_EXACT | | |
814 | RQFCR_HASHTBL_0 | RQFCR_PID_MASK; | |
815 | ftp_rqfpr[l] = FPR_FILER_MASK; | |
816 | gfar_write_filer(priv, l, ftp_rqfcr[l], ftp_rqfpr[l]); | |
817 | break; | |
818 | } | |
819 | ||
820 | if (!(ftp_rqfcr[l] & RQFCR_CLE) && (ftp_rqfcr[l] & RQFCR_AND)) | |
821 | continue; | |
822 | else { | |
823 | local_rqfpr[j] = ftp_rqfpr[l]; | |
824 | local_rqfcr[j] = ftp_rqfcr[l]; | |
825 | j--; | |
826 | } | |
827 | } | |
828 | ||
829 | priv->cur_filer_idx = l - 1; | |
830 | last_rule_idx = l; | |
831 | ||
832 | /* hash rules */ | |
833 | ethflow_to_filer_rules(priv, ethflow); | |
834 | ||
835 | /* Write back the popped out rules again */ | |
836 | for (k = j+1; k < MAX_FILER_IDX; k++) { | |
837 | ftp_rqfpr[priv->cur_filer_idx] = local_rqfpr[k]; | |
838 | ftp_rqfcr[priv->cur_filer_idx] = local_rqfcr[k]; | |
839 | gfar_write_filer(priv, priv->cur_filer_idx, | |
840 | local_rqfcr[k], local_rqfpr[k]); | |
841 | if (!priv->cur_filer_idx) | |
842 | break; | |
843 | priv->cur_filer_idx = priv->cur_filer_idx - 1; | |
844 | } | |
845 | ||
846 | return 1; | |
847 | } | |
848 | ||
849 | static int gfar_set_hash_opts(struct gfar_private *priv, struct ethtool_rxnfc *cmd) | |
850 | { | |
851 | u64 class; | |
852 | ||
853 | if (!gfar_ethflow_to_class(cmd->flow_type, &class)) | |
854 | return -EINVAL; | |
855 | ||
856 | if (class < CLASS_CODE_USER_PROG1 || | |
857 | class > CLASS_CODE_SCTP_IPV6) | |
858 | return -EINVAL; | |
859 | ||
860 | /* write the filer rules here */ | |
861 | if (!gfar_ethflow_to_filer_table(priv, cmd->data, cmd->flow_type)) | |
862 | return -1; | |
863 | ||
864 | return 0; | |
865 | } | |
866 | ||
867 | static int gfar_set_nfc(struct net_device *dev, struct ethtool_rxnfc *cmd) | |
868 | { | |
869 | struct gfar_private *priv = netdev_priv(dev); | |
870 | int ret = 0; | |
871 | ||
872 | switch(cmd->cmd) { | |
873 | case ETHTOOL_SRXFH: | |
874 | ret = gfar_set_hash_opts(priv, cmd); | |
875 | break; | |
876 | default: | |
877 | ret = -EINVAL; | |
878 | } | |
879 | ||
880 | return ret; | |
881 | } | |
882 | ||
7282d491 | 883 | const struct ethtool_ops gfar_ethtool_ops = { |
1da177e4 | 884 | .get_settings = gfar_gsettings, |
bb40dcbb | 885 | .set_settings = gfar_ssettings, |
1da177e4 LT |
886 | .get_drvinfo = gfar_gdrvinfo, |
887 | .get_regs_len = gfar_reglen, | |
888 | .get_regs = gfar_get_regs, | |
889 | .get_link = ethtool_op_get_link, | |
890 | .get_coalesce = gfar_gcoalesce, | |
891 | .set_coalesce = gfar_scoalesce, | |
892 | .get_ringparam = gfar_gringparam, | |
893 | .set_ringparam = gfar_sringparam, | |
894 | .get_strings = gfar_gstrings, | |
b9f2c044 | 895 | .get_sset_count = gfar_sset_count, |
1da177e4 | 896 | .get_ethtool_stats = gfar_fill_stats, |
0bbaf069 KG |
897 | .get_rx_csum = gfar_get_rx_csum, |
898 | .get_tx_csum = gfar_get_tx_csum, | |
899 | .set_rx_csum = gfar_set_rx_csum, | |
900 | .set_tx_csum = gfar_set_tx_csum, | |
4669bc90 | 901 | .set_sg = ethtool_op_set_sg, |
0bbaf069 KG |
902 | .get_msglevel = gfar_get_msglevel, |
903 | .set_msglevel = gfar_set_msglevel, | |
d87eb127 SW |
904 | #ifdef CONFIG_PM |
905 | .get_wol = gfar_get_wol, | |
906 | .set_wol = gfar_set_wol, | |
907 | #endif | |
7a8b3372 | 908 | .set_rxnfc = gfar_set_nfc, |
1da177e4 | 909 | }; |