SUNRPC: Allow cache lookups to use RCU protection rather than the r/w spinlock
[linux-2.6-block.git] / net / dsa / dsa.c
CommitLineData
91da11f8
LB
1/*
2 * net/dsa/dsa.c - Hardware switch handling
e84665c9 3 * Copyright (c) 2008-2009 Marvell Semiconductor
5e95329b 4 * Copyright (c) 2013 Florian Fainelli <florian@openwrt.org>
91da11f8
LB
5 *
6 * This program is free software; you can redistribute it and/or modify
7 * it under the terms of the GNU General Public License as published by
8 * the Free Software Foundation; either version 2 of the License, or
9 * (at your option) any later version.
10 */
11
51579c3f 12#include <linux/device.h>
91da11f8 13#include <linux/list.h>
91da11f8 14#include <linux/platform_device.h>
5a0e3ad6 15#include <linux/slab.h>
3a9a231d 16#include <linux/module.h>
60724d4b 17#include <linux/notifier.h>
5e95329b
FF
18#include <linux/of.h>
19#include <linux/of_mdio.h>
20#include <linux/of_platform.h>
769a0202 21#include <linux/of_net.h>
c6e970a0 22#include <linux/netdevice.h>
51579c3f 23#include <linux/sysfs.h>
cbc5d90b 24#include <linux/phy_fixed.h>
90af1059 25#include <linux/ptp_classify.h>
a86d8bec 26#include <linux/etherdevice.h>
ea5dd34b 27
91da11f8
LB
28#include "dsa_priv.h"
29
39a7f2a4
AL
30static struct sk_buff *dsa_slave_notag_xmit(struct sk_buff *skb,
31 struct net_device *dev)
32{
33 /* Just return the original SKB */
34 return skb;
35}
36
37static const struct dsa_device_ops none_ops = {
38 .xmit = dsa_slave_notag_xmit,
39 .rcv = NULL,
40};
41
42const struct dsa_device_ops *dsa_device_ops[DSA_TAG_LAST] = {
eb7b7211
AL
43#ifdef CONFIG_NET_DSA_TAG_BRCM
44 [DSA_TAG_PROTO_BRCM] = &brcm_netdev_ops,
45#endif
b74b70c4
FF
46#ifdef CONFIG_NET_DSA_TAG_BRCM_PREPEND
47 [DSA_TAG_PROTO_BRCM_PREPEND] = &brcm_prepend_netdev_ops,
48#endif
39a7f2a4
AL
49#ifdef CONFIG_NET_DSA_TAG_DSA
50 [DSA_TAG_PROTO_DSA] = &dsa_netdev_ops,
51#endif
52#ifdef CONFIG_NET_DSA_TAG_EDSA
53 [DSA_TAG_PROTO_EDSA] = &edsa_netdev_ops,
54#endif
8b8010fb
WH
55#ifdef CONFIG_NET_DSA_TAG_KSZ
56 [DSA_TAG_PROTO_KSZ] = &ksz_netdev_ops,
57#endif
eb7b7211
AL
58#ifdef CONFIG_NET_DSA_TAG_LAN9303
59 [DSA_TAG_PROTO_LAN9303] = &lan9303_netdev_ops,
39a7f2a4 60#endif
eb7b7211
AL
61#ifdef CONFIG_NET_DSA_TAG_MTK
62 [DSA_TAG_PROTO_MTK] = &mtk_netdev_ops,
cafdc45c
JC
63#endif
64#ifdef CONFIG_NET_DSA_TAG_QCA
65 [DSA_TAG_PROTO_QCA] = &qca_netdev_ops,
5cd8985a 66#endif
eb7b7211
AL
67#ifdef CONFIG_NET_DSA_TAG_TRAILER
68 [DSA_TAG_PROTO_TRAILER] = &trailer_netdev_ops,
39a7f2a4
AL
69#endif
70 [DSA_TAG_PROTO_NONE] = &none_ops,
71};
91da11f8 72
39a7f2a4
AL
73const struct dsa_device_ops *dsa_resolve_tag_protocol(int tag_protocol)
74{
75 const struct dsa_device_ops *ops;
76
77 if (tag_protocol >= DSA_TAG_LAST)
78 return ERR_PTR(-EINVAL);
79 ops = dsa_device_ops[tag_protocol];
80
81 if (!ops)
82 return ERR_PTR(-ENOPROTOOPT);
83
84 return ops;
85}
86
91da11f8
LB
87static int dev_is_class(struct device *dev, void *class)
88{
89 if (dev->class != NULL && !strcmp(dev->class->name, class))
90 return 1;
91
92 return 0;
93}
94
95static struct device *dev_find_class(struct device *parent, char *class)
96{
97 if (dev_is_class(parent, class)) {
98 get_device(parent);
99 return parent;
100 }
101
102 return device_find_child(parent, class, dev_is_class);
103}
104
14b89f36 105struct net_device *dsa_dev_to_net_device(struct device *dev)
91da11f8
LB
106{
107 struct device *d;
108
109 d = dev_find_class(dev, "net");
110 if (d != NULL) {
111 struct net_device *nd;
112
113 nd = to_net_dev(d);
114 dev_hold(nd);
115 put_device(d);
116
117 return nd;
118 }
119
120 return NULL;
121}
14b89f36 122EXPORT_SYMBOL_GPL(dsa_dev_to_net_device);
91da11f8 123
90af1059
BS
124/* Determine if we should defer delivery of skb until we have a rx timestamp.
125 *
126 * Called from dsa_switch_rcv. For now, this will only work if tagging is
127 * enabled on the switch. Normally the MAC driver would retrieve the hardware
128 * timestamp when it reads the packet out of the hardware. However in a DSA
129 * switch, the DSA driver owning the interface to which the packet is
130 * delivered is never notified unless we do so here.
131 */
132static bool dsa_skb_defer_rx_timestamp(struct dsa_slave_priv *p,
133 struct sk_buff *skb)
134{
135 struct dsa_switch *ds = p->dp->ds;
136 unsigned int type;
137
138 if (skb_headroom(skb) < ETH_HLEN)
139 return false;
140
141 __skb_push(skb, ETH_HLEN);
142
143 type = ptp_classify_raw(skb);
144
145 __skb_pull(skb, ETH_HLEN);
146
147 if (type == PTP_CLASS_NONE)
148 return false;
149
150 if (likely(ds->ops->port_rxtstamp))
151 return ds->ops->port_rxtstamp(ds, p->dp->index, skb, type);
152
153 return false;
154}
155
3e8a72d1 156static int dsa_switch_rcv(struct sk_buff *skb, struct net_device *dev,
89e49506 157 struct packet_type *pt, struct net_device *unused)
3e8a72d1 158{
2f657a60 159 struct dsa_port *cpu_dp = dev->dsa_ptr;
a86d8bec 160 struct sk_buff *nskb = NULL;
5f6b4e14 161 struct pcpu_sw_netstats *s;
f613ed66 162 struct dsa_slave_priv *p;
3e8a72d1 163
2f657a60 164 if (unlikely(!cpu_dp)) {
3e8a72d1
FF
165 kfree_skb(skb);
166 return 0;
167 }
168
16c5dcb1
FF
169 skb = skb_unshare(skb, GFP_ATOMIC);
170 if (!skb)
171 return 0;
172
2f657a60 173 nskb = cpu_dp->rcv(skb, dev, pt);
a86d8bec
FF
174 if (!nskb) {
175 kfree_skb(skb);
176 return 0;
177 }
178
179 skb = nskb;
f613ed66 180 p = netdev_priv(skb->dev);
a86d8bec
FF
181 skb_push(skb, ETH_HLEN);
182 skb->pkt_type = PACKET_HOST;
183 skb->protocol = eth_type_trans(skb, skb->dev);
184
5f6b4e14
FF
185 s = this_cpu_ptr(p->stats64);
186 u64_stats_update_begin(&s->syncp);
187 s->rx_packets++;
188 s->rx_bytes += skb->len;
189 u64_stats_update_end(&s->syncp);
a86d8bec 190
90af1059
BS
191 if (dsa_skb_defer_rx_timestamp(p, skb))
192 return 0;
193
a86d8bec
FF
194 netif_receive_skb(skb);
195
196 return 0;
3e8a72d1
FF
197}
198
ac2629a4 199#ifdef CONFIG_PM_SLEEP
e7d53ad3
VD
200static bool dsa_is_port_initialized(struct dsa_switch *ds, int p)
201{
4a5b85ff 202 return dsa_is_user_port(ds, p) && ds->ports[p].slave;
e7d53ad3
VD
203}
204
ac2629a4
FF
205int dsa_switch_suspend(struct dsa_switch *ds)
206{
207 int i, ret = 0;
208
209 /* Suspend slave network devices */
210 for (i = 0; i < ds->num_ports; i++) {
211 if (!dsa_is_port_initialized(ds, i))
212 continue;
213
f8b8b1cd 214 ret = dsa_slave_suspend(ds->ports[i].slave);
ac2629a4
FF
215 if (ret)
216 return ret;
217 }
218
219 if (ds->ops->suspend)
220 ret = ds->ops->suspend(ds);
221
222 return ret;
223}
224EXPORT_SYMBOL_GPL(dsa_switch_suspend);
225
226int dsa_switch_resume(struct dsa_switch *ds)
227{
228 int i, ret = 0;
229
230 if (ds->ops->resume)
231 ret = ds->ops->resume(ds);
232
233 if (ret)
234 return ret;
235
236 /* Resume slave network devices */
237 for (i = 0; i < ds->num_ports; i++) {
238 if (!dsa_is_port_initialized(ds, i))
239 continue;
240
f8b8b1cd 241 ret = dsa_slave_resume(ds->ports[i].slave);
ac2629a4
FF
242 if (ret)
243 return ret;
244 }
245
246 return 0;
247}
248EXPORT_SYMBOL_GPL(dsa_switch_resume);
249#endif
250
61b7363f 251static struct packet_type dsa_pack_type __read_mostly = {
3e8a72d1
FF
252 .type = cpu_to_be16(ETH_P_XDSA),
253 .func = dsa_switch_rcv,
254};
255
c9eb3e0f
AS
256static struct workqueue_struct *dsa_owq;
257
258bool dsa_schedule_work(struct work_struct *work)
259{
260 return queue_work(dsa_owq, work);
261}
262
60724d4b
FF
263static ATOMIC_NOTIFIER_HEAD(dsa_notif_chain);
264
265int register_dsa_notifier(struct notifier_block *nb)
266{
267 return atomic_notifier_chain_register(&dsa_notif_chain, nb);
268}
269EXPORT_SYMBOL_GPL(register_dsa_notifier);
270
271int unregister_dsa_notifier(struct notifier_block *nb)
272{
273 return atomic_notifier_chain_unregister(&dsa_notif_chain, nb);
274}
275EXPORT_SYMBOL_GPL(unregister_dsa_notifier);
276
277int call_dsa_notifiers(unsigned long val, struct net_device *dev,
278 struct dsa_notifier_info *info)
279{
280 info->dev = dev;
281 return atomic_notifier_call_chain(&dsa_notif_chain, val, info);
282}
283EXPORT_SYMBOL_GPL(call_dsa_notifiers);
284
91da11f8
LB
285static int __init dsa_init_module(void)
286{
7df899c3
BH
287 int rc;
288
c9eb3e0f
AS
289 dsa_owq = alloc_ordered_workqueue("dsa_ordered",
290 WQ_MEM_RECLAIM);
291 if (!dsa_owq)
292 return -ENOMEM;
293
88e4f0ca
VD
294 rc = dsa_slave_register_notifier();
295 if (rc)
296 return rc;
b73adef6 297
a6a71f19 298 rc = dsa_legacy_register();
7df899c3
BH
299 if (rc)
300 return rc;
301
3e8a72d1
FF
302 dev_add_pack(&dsa_pack_type);
303
7df899c3 304 return 0;
91da11f8
LB
305}
306module_init(dsa_init_module);
307
308static void __exit dsa_cleanup_module(void)
309{
88e4f0ca 310 dsa_slave_unregister_notifier();
3e8a72d1 311 dev_remove_pack(&dsa_pack_type);
a6a71f19 312 dsa_legacy_unregister();
c9eb3e0f 313 destroy_workqueue(dsa_owq);
91da11f8
LB
314}
315module_exit(dsa_cleanup_module);
316
577d6a7c 317MODULE_AUTHOR("Lennert Buytenhek <buytenh@wantstofly.org>");
91da11f8
LB
318MODULE_DESCRIPTION("Driver for Distributed Switch Architecture switch chips");
319MODULE_LICENSE("GPL");
320MODULE_ALIAS("platform:dsa");