Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | #ifndef __MV643XX_ETH_H__ |
2 | #define __MV643XX_ETH_H__ | |
3 | ||
1da177e4 LT |
4 | #include <linux/module.h> |
5 | #include <linux/kernel.h> | |
6 | #include <linux/spinlock.h> | |
7 | #include <linux/workqueue.h> | |
c28a4f89 | 8 | #include <linux/mii.h> |
1da177e4 LT |
9 | |
10 | #include <linux/mv643xx.h> | |
11 | ||
471a5671 DJ |
12 | #include <asm/dma-mapping.h> |
13 | ||
1da177e4 LT |
14 | /* Checksum offload for Tx works for most packets, but |
15 | * fails if previous packet sent did not use hw csum | |
16 | */ | |
26006360 | 17 | #define MV643XX_CHECKSUM_OFFLOAD_TX |
1da177e4 LT |
18 | #define MV643XX_NAPI |
19 | #define MV643XX_TX_FAST_REFILL | |
1da177e4 LT |
20 | #undef MV643XX_COAL |
21 | ||
22 | /* | |
23 | * Number of RX / TX descriptors on RX / TX rings. | |
24 | * Note that allocating RX descriptors is done by allocating the RX | |
25 | * ring AND a preallocated RX buffers (skb's) for each descriptor. | |
26 | * The TX descriptors only allocates the TX descriptors ring, | |
27 | * with no pre allocated TX buffers (skb's are allocated by higher layers. | |
28 | */ | |
29 | ||
30 | /* Default TX ring size is 1000 descriptors */ | |
31 | #define MV643XX_DEFAULT_TX_QUEUE_SIZE 1000 | |
32 | ||
33 | /* Default RX ring size is 400 descriptors */ | |
34 | #define MV643XX_DEFAULT_RX_QUEUE_SIZE 400 | |
35 | ||
36 | #define MV643XX_TX_COAL 100 | |
37 | #ifdef MV643XX_COAL | |
38 | #define MV643XX_RX_COAL 100 | |
39 | #endif | |
40 | ||
7303fde8 DF |
41 | #ifdef MV643XX_CHECKSUM_OFFLOAD_TX |
42 | #define MAX_DESCS_PER_SKB (MAX_SKB_FRAGS + 1) | |
43 | #else | |
44 | #define MAX_DESCS_PER_SKB 1 | |
45 | #endif | |
1da177e4 | 46 | |
7303fde8 DF |
47 | #define ETH_VLAN_HLEN 4 |
48 | #define ETH_FCS_LEN 4 | |
6f059c3e | 49 | #define ETH_HW_IP_ALIGN 2 /* hw aligns IP header */ |
7303fde8 | 50 | #define ETH_WRAPPER_LEN (ETH_HW_IP_ALIGN + ETH_HLEN + \ |
6f059c3e | 51 | ETH_VLAN_HLEN + ETH_FCS_LEN) |
471a5671 | 52 | #define ETH_RX_SKB_SIZE (dev->mtu + ETH_WRAPPER_LEN + dma_get_cache_alignment()) |
7303fde8 DF |
53 | |
54 | #define ETH_RX_QUEUES_ENABLED (1 << 0) /* use only Q0 for receive */ | |
55 | #define ETH_TX_QUEUES_ENABLED (1 << 0) /* use only Q0 for transmit */ | |
56 | ||
57 | #define ETH_INT_CAUSE_RX_DONE (ETH_RX_QUEUES_ENABLED << 2) | |
58 | #define ETH_INT_CAUSE_RX_ERROR (ETH_RX_QUEUES_ENABLED << 9) | |
59 | #define ETH_INT_CAUSE_RX (ETH_INT_CAUSE_RX_DONE | ETH_INT_CAUSE_RX_ERROR) | |
60 | #define ETH_INT_CAUSE_EXT 0x00000002 | |
61 | #define ETH_INT_UNMASK_ALL (ETH_INT_CAUSE_RX | ETH_INT_CAUSE_EXT) | |
62 | ||
63 | #define ETH_INT_CAUSE_TX_DONE (ETH_TX_QUEUES_ENABLED << 0) | |
64 | #define ETH_INT_CAUSE_TX_ERROR (ETH_TX_QUEUES_ENABLED << 8) | |
65 | #define ETH_INT_CAUSE_TX (ETH_INT_CAUSE_TX_DONE | ETH_INT_CAUSE_TX_ERROR) | |
66 | #define ETH_INT_CAUSE_PHY 0x00010000 | |
67 | #define ETH_INT_UNMASK_ALL_EXT (ETH_INT_CAUSE_TX | ETH_INT_CAUSE_PHY) | |
68 | ||
69 | #define ETH_INT_MASK_ALL 0x00000000 | |
70 | #define ETH_INT_MASK_ALL_EXT 0x00000000 | |
71 | ||
72 | #define PHY_WAIT_ITERATIONS 1000 /* 1000 iterations * 10uS = 10mS max */ | |
73 | #define PHY_WAIT_MICRO_SECONDS 10 | |
1da177e4 | 74 | |
1da177e4 LT |
75 | /* Buffer offset from buffer pointer */ |
76 | #define RX_BUF_OFFSET 0x2 | |
77 | ||
78 | /* Gigabit Ethernet Unit Global Registers */ | |
79 | ||
80 | /* MIB Counters register definitions */ | |
81 | #define ETH_MIB_GOOD_OCTETS_RECEIVED_LOW 0x0 | |
82 | #define ETH_MIB_GOOD_OCTETS_RECEIVED_HIGH 0x4 | |
83 | #define ETH_MIB_BAD_OCTETS_RECEIVED 0x8 | |
84 | #define ETH_MIB_INTERNAL_MAC_TRANSMIT_ERR 0xc | |
85 | #define ETH_MIB_GOOD_FRAMES_RECEIVED 0x10 | |
86 | #define ETH_MIB_BAD_FRAMES_RECEIVED 0x14 | |
87 | #define ETH_MIB_BROADCAST_FRAMES_RECEIVED 0x18 | |
88 | #define ETH_MIB_MULTICAST_FRAMES_RECEIVED 0x1c | |
89 | #define ETH_MIB_FRAMES_64_OCTETS 0x20 | |
90 | #define ETH_MIB_FRAMES_65_TO_127_OCTETS 0x24 | |
91 | #define ETH_MIB_FRAMES_128_TO_255_OCTETS 0x28 | |
92 | #define ETH_MIB_FRAMES_256_TO_511_OCTETS 0x2c | |
93 | #define ETH_MIB_FRAMES_512_TO_1023_OCTETS 0x30 | |
94 | #define ETH_MIB_FRAMES_1024_TO_MAX_OCTETS 0x34 | |
95 | #define ETH_MIB_GOOD_OCTETS_SENT_LOW 0x38 | |
96 | #define ETH_MIB_GOOD_OCTETS_SENT_HIGH 0x3c | |
97 | #define ETH_MIB_GOOD_FRAMES_SENT 0x40 | |
98 | #define ETH_MIB_EXCESSIVE_COLLISION 0x44 | |
99 | #define ETH_MIB_MULTICAST_FRAMES_SENT 0x48 | |
100 | #define ETH_MIB_BROADCAST_FRAMES_SENT 0x4c | |
101 | #define ETH_MIB_UNREC_MAC_CONTROL_RECEIVED 0x50 | |
102 | #define ETH_MIB_FC_SENT 0x54 | |
103 | #define ETH_MIB_GOOD_FC_RECEIVED 0x58 | |
104 | #define ETH_MIB_BAD_FC_RECEIVED 0x5c | |
105 | #define ETH_MIB_UNDERSIZE_RECEIVED 0x60 | |
106 | #define ETH_MIB_FRAGMENTS_RECEIVED 0x64 | |
107 | #define ETH_MIB_OVERSIZE_RECEIVED 0x68 | |
108 | #define ETH_MIB_JABBER_RECEIVED 0x6c | |
109 | #define ETH_MIB_MAC_RECEIVE_ERROR 0x70 | |
110 | #define ETH_MIB_BAD_CRC_EVENT 0x74 | |
111 | #define ETH_MIB_COLLISION 0x78 | |
112 | #define ETH_MIB_LATE_COLLISION 0x7c | |
113 | ||
114 | /* Port serial status reg (PSR) */ | |
ebe19a4e DF |
115 | #define ETH_INTERFACE_PCM 0x00000001 |
116 | #define ETH_LINK_IS_UP 0x00000002 | |
117 | #define ETH_PORT_AT_FULL_DUPLEX 0x00000004 | |
118 | #define ETH_RX_FLOW_CTRL_ENABLED 0x00000008 | |
119 | #define ETH_GMII_SPEED_1000 0x00000010 | |
120 | #define ETH_MII_SPEED_100 0x00000020 | |
121 | #define ETH_TX_IN_PROGRESS 0x00000080 | |
122 | #define ETH_BYPASS_ACTIVE 0x00000100 | |
123 | #define ETH_PORT_AT_PARTITION_STATE 0x00000200 | |
124 | #define ETH_PORT_TX_FIFO_EMPTY 0x00000400 | |
1da177e4 LT |
125 | |
126 | /* SMI reg */ | |
ebe19a4e DF |
127 | #define ETH_SMI_BUSY 0x10000000 /* 0 - Write, 1 - Read */ |
128 | #define ETH_SMI_READ_VALID 0x08000000 /* 0 - Write, 1 - Read */ | |
129 | #define ETH_SMI_OPCODE_WRITE 0 /* Completion of Read */ | |
130 | #define ETH_SMI_OPCODE_READ 0x04000000 /* Operation is in progress */ | |
131 | ||
132 | /* Interrupt Cause Register Bit Definitions */ | |
1da177e4 LT |
133 | |
134 | /* SDMA command status fields macros */ | |
135 | ||
136 | /* Tx & Rx descriptors status */ | |
ebe19a4e | 137 | #define ETH_ERROR_SUMMARY 0x00000001 |
1da177e4 LT |
138 | |
139 | /* Tx & Rx descriptors command */ | |
ebe19a4e | 140 | #define ETH_BUFFER_OWNED_BY_DMA 0x80000000 |
1da177e4 LT |
141 | |
142 | /* Tx descriptors status */ | |
ebe19a4e DF |
143 | #define ETH_LC_ERROR 0 |
144 | #define ETH_UR_ERROR 0x00000002 | |
145 | #define ETH_RL_ERROR 0x00000004 | |
146 | #define ETH_LLC_SNAP_FORMAT 0x00000200 | |
1da177e4 LT |
147 | |
148 | /* Rx descriptors status */ | |
ebe19a4e DF |
149 | #define ETH_OVERRUN_ERROR 0x00000002 |
150 | #define ETH_MAX_FRAME_LENGTH_ERROR 0x00000004 | |
151 | #define ETH_RESOURCE_ERROR 0x00000006 | |
152 | #define ETH_VLAN_TAGGED 0x00080000 | |
153 | #define ETH_BPDU_FRAME 0x00100000 | |
154 | #define ETH_UDP_FRAME_OVER_IP_V_4 0x00200000 | |
155 | #define ETH_OTHER_FRAME_TYPE 0x00400000 | |
156 | #define ETH_LAYER_2_IS_ETH_V_2 0x00800000 | |
157 | #define ETH_FRAME_TYPE_IP_V_4 0x01000000 | |
158 | #define ETH_FRAME_HEADER_OK 0x02000000 | |
159 | #define ETH_RX_LAST_DESC 0x04000000 | |
160 | #define ETH_RX_FIRST_DESC 0x08000000 | |
161 | #define ETH_UNKNOWN_DESTINATION_ADDR 0x10000000 | |
162 | #define ETH_RX_ENABLE_INTERRUPT 0x20000000 | |
163 | #define ETH_LAYER_4_CHECKSUM_OK 0x40000000 | |
1da177e4 LT |
164 | |
165 | /* Rx descriptors byte count */ | |
ebe19a4e | 166 | #define ETH_FRAME_FRAGMENTED 0x00000004 |
1da177e4 LT |
167 | |
168 | /* Tx descriptors command */ | |
ebe19a4e DF |
169 | #define ETH_LAYER_4_CHECKSUM_FIRST_DESC 0x00000400 |
170 | #define ETH_FRAME_SET_TO_VLAN 0x00008000 | |
171 | #define ETH_UDP_FRAME 0x00010000 | |
172 | #define ETH_GEN_TCP_UDP_CHECKSUM 0x00020000 | |
173 | #define ETH_GEN_IP_V_4_CHECKSUM 0x00040000 | |
174 | #define ETH_ZERO_PADDING 0x00080000 | |
175 | #define ETH_TX_LAST_DESC 0x00100000 | |
176 | #define ETH_TX_FIRST_DESC 0x00200000 | |
177 | #define ETH_GEN_CRC 0x00400000 | |
178 | #define ETH_TX_ENABLE_INTERRUPT 0x00800000 | |
179 | #define ETH_AUTO_MODE 0x40000000 | |
1da177e4 | 180 | |
26006360 DF |
181 | #define ETH_TX_IHL_SHIFT 11 |
182 | ||
1da177e4 LT |
183 | /* typedefs */ |
184 | ||
185 | typedef enum _eth_func_ret_status { | |
186 | ETH_OK, /* Returned as expected. */ | |
187 | ETH_ERROR, /* Fundamental error. */ | |
188 | ETH_RETRY, /* Could not process request. Try later.*/ | |
189 | ETH_END_OF_JOB, /* Ring has nothing to process. */ | |
190 | ETH_QUEUE_FULL, /* Ring resource error. */ | |
191 | ETH_QUEUE_LAST_RESOURCE /* Ring resources about to exhaust. */ | |
192 | } ETH_FUNC_RET_STATUS; | |
193 | ||
194 | typedef enum _eth_target { | |
195 | ETH_TARGET_DRAM, | |
196 | ETH_TARGET_DEVICE, | |
197 | ETH_TARGET_CBS, | |
198 | ETH_TARGET_PCI0, | |
199 | ETH_TARGET_PCI1 | |
200 | } ETH_TARGET; | |
201 | ||
202 | /* These are for big-endian machines. Little endian needs different | |
203 | * definitions. | |
204 | */ | |
205 | #if defined(__BIG_ENDIAN) | |
206 | struct eth_rx_desc { | |
207 | u16 byte_cnt; /* Descriptor buffer byte count */ | |
208 | u16 buf_size; /* Buffer size */ | |
209 | u32 cmd_sts; /* Descriptor command status */ | |
210 | u32 next_desc_ptr; /* Next descriptor pointer */ | |
211 | u32 buf_ptr; /* Descriptor buffer pointer */ | |
212 | }; | |
213 | ||
214 | struct eth_tx_desc { | |
215 | u16 byte_cnt; /* buffer byte count */ | |
216 | u16 l4i_chk; /* CPU provided TCP checksum */ | |
217 | u32 cmd_sts; /* Command/status field */ | |
218 | u32 next_desc_ptr; /* Pointer to next descriptor */ | |
219 | u32 buf_ptr; /* pointer to buffer for this descriptor*/ | |
220 | }; | |
221 | ||
222 | #elif defined(__LITTLE_ENDIAN) | |
223 | struct eth_rx_desc { | |
224 | u32 cmd_sts; /* Descriptor command status */ | |
225 | u16 buf_size; /* Buffer size */ | |
226 | u16 byte_cnt; /* Descriptor buffer byte count */ | |
227 | u32 buf_ptr; /* Descriptor buffer pointer */ | |
228 | u32 next_desc_ptr; /* Next descriptor pointer */ | |
229 | }; | |
230 | ||
231 | struct eth_tx_desc { | |
232 | u32 cmd_sts; /* Command/status field */ | |
233 | u16 l4i_chk; /* CPU provided TCP checksum */ | |
234 | u16 byte_cnt; /* buffer byte count */ | |
235 | u32 buf_ptr; /* pointer to buffer for this descriptor*/ | |
236 | u32 next_desc_ptr; /* Pointer to next descriptor */ | |
237 | }; | |
238 | #else | |
239 | #error One of __BIG_ENDIAN or __LITTLE_ENDIAN must be defined | |
240 | #endif | |
241 | ||
242 | /* Unified struct for Rx and Tx operations. The user is not required to */ | |
243 | /* be familier with neither Tx nor Rx descriptors. */ | |
244 | struct pkt_info { | |
245 | unsigned short byte_cnt; /* Descriptor buffer byte count */ | |
246 | unsigned short l4i_chk; /* Tx CPU provided TCP Checksum */ | |
247 | unsigned int cmd_sts; /* Descriptor command status */ | |
248 | dma_addr_t buf_ptr; /* Descriptor buffer pointer */ | |
249 | struct sk_buff *return_info; /* User resource return information */ | |
250 | }; | |
251 | ||
47bdd718 | 252 | /* Ethernet port specific information */ |
1da177e4 LT |
253 | |
254 | struct mv643xx_mib_counters { | |
255 | u64 good_octets_received; | |
256 | u32 bad_octets_received; | |
257 | u32 internal_mac_transmit_err; | |
258 | u32 good_frames_received; | |
259 | u32 bad_frames_received; | |
260 | u32 broadcast_frames_received; | |
261 | u32 multicast_frames_received; | |
262 | u32 frames_64_octets; | |
263 | u32 frames_65_to_127_octets; | |
264 | u32 frames_128_to_255_octets; | |
265 | u32 frames_256_to_511_octets; | |
266 | u32 frames_512_to_1023_octets; | |
267 | u32 frames_1024_to_max_octets; | |
268 | u64 good_octets_sent; | |
269 | u32 good_frames_sent; | |
270 | u32 excessive_collision; | |
271 | u32 multicast_frames_sent; | |
272 | u32 broadcast_frames_sent; | |
273 | u32 unrec_mac_control_received; | |
274 | u32 fc_sent; | |
275 | u32 good_fc_received; | |
276 | u32 bad_fc_received; | |
277 | u32 undersize_received; | |
278 | u32 fragments_received; | |
279 | u32 oversize_received; | |
280 | u32 jabber_received; | |
281 | u32 mac_receive_error; | |
282 | u32 bad_crc_event; | |
283 | u32 collision; | |
284 | u32 late_collision; | |
285 | }; | |
286 | ||
287 | struct mv643xx_private { | |
288 | int port_num; /* User Ethernet port number */ | |
1da177e4 LT |
289 | |
290 | u32 rx_sram_addr; /* Base address of rx sram area */ | |
291 | u32 rx_sram_size; /* Size of rx sram area */ | |
292 | u32 tx_sram_addr; /* Base address of tx sram area */ | |
293 | u32 tx_sram_size; /* Size of tx sram area */ | |
294 | ||
295 | int rx_resource_err; /* Rx ring resource error flag */ | |
1da177e4 LT |
296 | |
297 | /* Tx/Rx rings managment indexes fields. For driver use */ | |
298 | ||
299 | /* Next available and first returning Rx resource */ | |
300 | int rx_curr_desc_q, rx_used_desc_q; | |
301 | ||
302 | /* Next available and first returning Tx resource */ | |
303 | int tx_curr_desc_q, tx_used_desc_q; | |
1da177e4 LT |
304 | |
305 | #ifdef MV643XX_TX_FAST_REFILL | |
306 | u32 tx_clean_threshold; | |
307 | #endif | |
308 | ||
309 | struct eth_rx_desc *p_rx_desc_area; | |
310 | dma_addr_t rx_desc_dma; | |
c8aaea25 | 311 | int rx_desc_area_size; |
1da177e4 LT |
312 | struct sk_buff **rx_skb; |
313 | ||
314 | struct eth_tx_desc *p_tx_desc_area; | |
315 | dma_addr_t tx_desc_dma; | |
c8aaea25 | 316 | int tx_desc_area_size; |
1da177e4 LT |
317 | struct sk_buff **tx_skb; |
318 | ||
319 | struct work_struct tx_timeout_task; | |
320 | ||
1da177e4 LT |
321 | struct net_device_stats stats; |
322 | struct mv643xx_mib_counters mib_counters; | |
323 | spinlock_t lock; | |
324 | /* Size of Tx Ring per queue */ | |
c8aaea25 | 325 | int tx_ring_size; |
f98e36f1 | 326 | /* Number of tx descriptors in use */ |
c8aaea25 | 327 | int tx_desc_count; |
1da177e4 | 328 | /* Size of Rx Ring per queue */ |
c8aaea25 | 329 | int rx_ring_size; |
f98e36f1 | 330 | /* Number of rx descriptors in use */ |
c8aaea25 | 331 | int rx_desc_count; |
1da177e4 | 332 | |
1da177e4 LT |
333 | /* |
334 | * Used in case RX Ring is empty, which can be caused when | |
335 | * system does not have resources (skb's) | |
336 | */ | |
337 | struct timer_list timeout; | |
1da177e4 LT |
338 | |
339 | u32 rx_int_coal; | |
340 | u32 tx_int_coal; | |
c28a4f89 | 341 | struct mii_if_info mii; |
1da177e4 LT |
342 | }; |
343 | ||
1da177e4 LT |
344 | /* Port operation control routines */ |
345 | static void eth_port_init(struct mv643xx_private *mp); | |
346 | static void eth_port_reset(unsigned int eth_port_num); | |
ed9b5d45 | 347 | static void eth_port_start(struct net_device *dev); |
1da177e4 | 348 | |
1da177e4 LT |
349 | /* Port MAC address routines */ |
350 | static void eth_port_uc_addr_set(unsigned int eth_port_num, | |
351 | unsigned char *p_addr); | |
352 | ||
353 | /* PHY and MIB routines */ | |
354 | static void ethernet_phy_reset(unsigned int eth_port_num); | |
355 | ||
356 | static void eth_port_write_smi_reg(unsigned int eth_port_num, | |
357 | unsigned int phy_reg, unsigned int value); | |
358 | ||
359 | static void eth_port_read_smi_reg(unsigned int eth_port_num, | |
360 | unsigned int phy_reg, unsigned int *value); | |
361 | ||
362 | static void eth_clear_mib_counters(unsigned int eth_port_num); | |
363 | ||
364 | /* Port data flow control routines */ | |
1da177e4 LT |
365 | static ETH_FUNC_RET_STATUS eth_port_receive(struct mv643xx_private *mp, |
366 | struct pkt_info *p_pkt_info); | |
367 | static ETH_FUNC_RET_STATUS eth_rx_return_buff(struct mv643xx_private *mp, | |
368 | struct pkt_info *p_pkt_info); | |
369 | ||
370 | #endif /* __MV643XX_ETH_H__ */ |