Commit | Line | Data |
---|---|---|
c72dfae2 SB |
1 | /* |
2 | * CAIF Interface registration. | |
3 | * Copyright (C) ST-Ericsson AB 2010 | |
4 | * Author: Sjur Brendeland/sjur.brandeland@stericsson.com | |
5 | * License terms: GNU General Public License (GPL) version 2 | |
6 | * | |
31fdc555 | 7 | * Borrowed heavily from file: pn_dev.c. Thanks to Remi Denis-Courmont |
c72dfae2 SB |
8 | * and Sakari Ailus <sakari.ailus@nokia.com> |
9 | */ | |
10 | ||
b31fa5ba JP |
11 | #define pr_fmt(fmt) KBUILD_MODNAME ":%s(): " fmt, __func__ |
12 | ||
c72dfae2 SB |
13 | #include <linux/kernel.h> |
14 | #include <linux/if_arp.h> | |
15 | #include <linux/net.h> | |
16 | #include <linux/netdevice.h> | |
bd30ce4b | 17 | #include <linux/mutex.h> |
3a9a231d | 18 | #include <linux/module.h> |
0e4c7d85 | 19 | #include <linux/spinlock.h> |
c72dfae2 SB |
20 | #include <net/netns/generic.h> |
21 | #include <net/net_namespace.h> | |
22 | #include <net/pkt_sched.h> | |
23 | #include <net/caif/caif_device.h> | |
c72dfae2 SB |
24 | #include <net/caif/caif_layer.h> |
25 | #include <net/caif/cfpkt.h> | |
26 | #include <net/caif/cfcnfg.h> | |
7c18d220 | 27 | #include <net/caif/cfserl.h> |
c72dfae2 SB |
28 | |
29 | MODULE_LICENSE("GPL"); | |
c72dfae2 SB |
30 | |
31 | /* Used for local tracking of the CAIF net devices */ | |
32 | struct caif_device_entry { | |
33 | struct cflayer layer; | |
34 | struct list_head list; | |
c72dfae2 | 35 | struct net_device *netdev; |
bd30ce4b | 36 | int __percpu *pcpu_refcnt; |
0e4c7d85 | 37 | spinlock_t flow_lock; |
7d311304 | 38 | struct sk_buff *xoff_skb; |
39 | void (*xoff_skb_dtor)(struct sk_buff *skb); | |
0e4c7d85 | 40 | bool xoff; |
c72dfae2 SB |
41 | }; |
42 | ||
43 | struct caif_device_entry_list { | |
44 | struct list_head list; | |
45 | /* Protects simulanous deletes in list */ | |
bd30ce4b | 46 | struct mutex lock; |
c72dfae2 SB |
47 | }; |
48 | ||
49 | struct caif_net { | |
bee925db | 50 | struct cfcnfg *cfg; |
c72dfae2 SB |
51 | struct caif_device_entry_list caifdevs; |
52 | }; | |
53 | ||
54 | static int caif_net_id; | |
0e4c7d85 | 55 | static int q_high = 50; /* Percent */ |
bee925db | 56 | |
57 | struct cfcnfg *get_cfcnfg(struct net *net) | |
58 | { | |
59 | struct caif_net *caifn; | |
bee925db | 60 | caifn = net_generic(net, caif_net_id); |
bee925db | 61 | return caifn->cfg; |
62 | } | |
63 | EXPORT_SYMBOL(get_cfcnfg); | |
c72dfae2 SB |
64 | |
65 | static struct caif_device_entry_list *caif_device_list(struct net *net) | |
66 | { | |
67 | struct caif_net *caifn; | |
c72dfae2 | 68 | caifn = net_generic(net, caif_net_id); |
c72dfae2 SB |
69 | return &caifn->caifdevs; |
70 | } | |
71 | ||
bd30ce4b | 72 | static void caifd_put(struct caif_device_entry *e) |
73 | { | |
933393f5 | 74 | this_cpu_dec(*e->pcpu_refcnt); |
bd30ce4b | 75 | } |
76 | ||
77 | static void caifd_hold(struct caif_device_entry *e) | |
78 | { | |
933393f5 | 79 | this_cpu_inc(*e->pcpu_refcnt); |
bd30ce4b | 80 | } |
81 | ||
82 | static int caifd_refcnt_read(struct caif_device_entry *e) | |
83 | { | |
84 | int i, refcnt = 0; | |
85 | for_each_possible_cpu(i) | |
86 | refcnt += *per_cpu_ptr(e->pcpu_refcnt, i); | |
87 | return refcnt; | |
88 | } | |
89 | ||
c72dfae2 SB |
90 | /* Allocate new CAIF device. */ |
91 | static struct caif_device_entry *caif_device_alloc(struct net_device *dev) | |
92 | { | |
c72dfae2 | 93 | struct caif_device_entry *caifd; |
bd30ce4b | 94 | |
4fb66b82 | 95 | caifd = kzalloc(sizeof(*caifd), GFP_KERNEL); |
c72dfae2 SB |
96 | if (!caifd) |
97 | return NULL; | |
bd30ce4b | 98 | caifd->pcpu_refcnt = alloc_percpu(int); |
4fb66b82 ED |
99 | if (!caifd->pcpu_refcnt) { |
100 | kfree(caifd); | |
101 | return NULL; | |
102 | } | |
c72dfae2 | 103 | caifd->netdev = dev; |
bd30ce4b | 104 | dev_hold(dev); |
c72dfae2 SB |
105 | return caifd; |
106 | } | |
107 | ||
108 | static struct caif_device_entry *caif_get(struct net_device *dev) | |
109 | { | |
110 | struct caif_device_entry_list *caifdevs = | |
111 | caif_device_list(dev_net(dev)); | |
112 | struct caif_device_entry *caifd; | |
7c18d220 | 113 | |
bd30ce4b | 114 | list_for_each_entry_rcu(caifd, &caifdevs->list, list) { |
c72dfae2 SB |
115 | if (caifd->netdev == dev) |
116 | return caifd; | |
117 | } | |
118 | return NULL; | |
119 | } | |
120 | ||
0e4c7d85 | 121 | void caif_flow_cb(struct sk_buff *skb) |
122 | { | |
123 | struct caif_device_entry *caifd; | |
7d311304 | 124 | void (*dtor)(struct sk_buff *skb) = NULL; |
0e4c7d85 | 125 | bool send_xoff; |
126 | ||
127 | WARN_ON(skb->dev == NULL); | |
128 | ||
129 | rcu_read_lock(); | |
130 | caifd = caif_get(skb->dev); | |
c95567c8 KLX |
131 | |
132 | WARN_ON(caifd == NULL); | |
133 | if (caifd == NULL) | |
134 | return; | |
135 | ||
0e4c7d85 | 136 | caifd_hold(caifd); |
137 | rcu_read_unlock(); | |
138 | ||
139 | spin_lock_bh(&caifd->flow_lock); | |
140 | send_xoff = caifd->xoff; | |
141 | caifd->xoff = 0; | |
59f608d8 | 142 | dtor = caifd->xoff_skb_dtor; |
143 | ||
144 | if (WARN_ON(caifd->xoff_skb != skb)) | |
145 | skb = NULL; | |
146 | ||
147 | caifd->xoff_skb = NULL; | |
148 | caifd->xoff_skb_dtor = NULL; | |
149 | ||
0e4c7d85 | 150 | spin_unlock_bh(&caifd->flow_lock); |
151 | ||
59f608d8 | 152 | if (dtor && skb) |
7d311304 | 153 | dtor(skb); |
154 | ||
0e4c7d85 | 155 | if (send_xoff) |
156 | caifd->layer.up-> | |
157 | ctrlcmd(caifd->layer.up, | |
158 | _CAIF_CTRLCMD_PHYIF_FLOW_ON_IND, | |
159 | caifd->layer.id); | |
160 | caifd_put(caifd); | |
161 | } | |
162 | ||
c72dfae2 SB |
163 | static int transmit(struct cflayer *layer, struct cfpkt *pkt) |
164 | { | |
0e4c7d85 | 165 | int err, high = 0, qlen = 0; |
c72dfae2 SB |
166 | struct caif_device_entry *caifd = |
167 | container_of(layer, struct caif_device_entry, layer); | |
4dd820c0 | 168 | struct sk_buff *skb; |
0e4c7d85 | 169 | struct netdev_queue *txq; |
170 | ||
171 | rcu_read_lock_bh(); | |
4dd820c0 | 172 | |
c72dfae2 SB |
173 | skb = cfpkt_tonative(pkt); |
174 | skb->dev = caifd->netdev; | |
7c18d220 | 175 | skb_reset_network_header(skb); |
176 | skb->protocol = htons(ETH_P_CAIF); | |
0e4c7d85 | 177 | |
178 | /* Check if we need to handle xoff */ | |
179 | if (likely(caifd->netdev->tx_queue_len == 0)) | |
180 | goto noxoff; | |
181 | ||
182 | if (unlikely(caifd->xoff)) | |
183 | goto noxoff; | |
184 | ||
185 | if (likely(!netif_queue_stopped(caifd->netdev))) { | |
186 | /* If we run with a TX queue, check if the queue is too long*/ | |
187 | txq = netdev_get_tx_queue(skb->dev, 0); | |
188 | qlen = qdisc_qlen(rcu_dereference_bh(txq->qdisc)); | |
189 | ||
190 | if (likely(qlen == 0)) | |
191 | goto noxoff; | |
192 | ||
193 | high = (caifd->netdev->tx_queue_len * q_high) / 100; | |
194 | if (likely(qlen < high)) | |
195 | goto noxoff; | |
196 | } | |
197 | ||
198 | /* Hold lock while accessing xoff */ | |
199 | spin_lock_bh(&caifd->flow_lock); | |
200 | if (caifd->xoff) { | |
201 | spin_unlock_bh(&caifd->flow_lock); | |
202 | goto noxoff; | |
203 | } | |
204 | ||
205 | /* | |
206 | * Handle flow off, we do this by temporary hi-jacking this | |
207 | * skb's destructor function, and replace it with our own | |
208 | * flow-on callback. The callback will set flow-on and call | |
209 | * the original destructor. | |
210 | */ | |
211 | ||
212 | pr_debug("queue has stopped(%d) or is full (%d > %d)\n", | |
213 | netif_queue_stopped(caifd->netdev), | |
214 | qlen, high); | |
215 | caifd->xoff = 1; | |
7d311304 | 216 | caifd->xoff_skb = skb; |
217 | caifd->xoff_skb_dtor = skb->destructor; | |
218 | skb->destructor = caif_flow_cb; | |
0e4c7d85 | 219 | spin_unlock_bh(&caifd->flow_lock); |
0e4c7d85 | 220 | |
221 | caifd->layer.up->ctrlcmd(caifd->layer.up, | |
222 | _CAIF_CTRLCMD_PHYIF_FLOW_OFF_IND, | |
223 | caifd->layer.id); | |
224 | noxoff: | |
225 | rcu_read_unlock_bh(); | |
4dd820c0 | 226 | |
c85c2951 | 227 | err = dev_queue_xmit(skb); |
228 | if (err > 0) | |
229 | err = -EIO; | |
c72dfae2 | 230 | |
c85c2951 | 231 | return err; |
c72dfae2 SB |
232 | } |
233 | ||
c72dfae2 | 234 | /* |
bd30ce4b | 235 | * Stuff received packets into the CAIF stack. |
c72dfae2 SB |
236 | * On error, returns non-zero and releases the skb. |
237 | */ | |
238 | static int receive(struct sk_buff *skb, struct net_device *dev, | |
239 | struct packet_type *pkttype, struct net_device *orig_dev) | |
240 | { | |
c72dfae2 SB |
241 | struct cfpkt *pkt; |
242 | struct caif_device_entry *caifd; | |
69c867c9 | 243 | int err; |
bd30ce4b | 244 | |
c72dfae2 | 245 | pkt = cfpkt_fromnative(CAIF_DIR_IN, skb); |
bd30ce4b | 246 | |
247 | rcu_read_lock(); | |
c72dfae2 | 248 | caifd = caif_get(dev); |
c72dfae2 | 249 | |
bd30ce4b | 250 | if (!caifd || !caifd->layer.up || !caifd->layer.up->receive || |
251 | !netif_oper_up(caifd->netdev)) { | |
252 | rcu_read_unlock(); | |
253 | kfree_skb(skb); | |
c72dfae2 | 254 | return NET_RX_DROP; |
bd30ce4b | 255 | } |
256 | ||
257 | /* Hold reference to netdevice while using CAIF stack */ | |
258 | caifd_hold(caifd); | |
259 | rcu_read_unlock(); | |
c72dfae2 | 260 | |
69c867c9 | 261 | err = caifd->layer.up->receive(caifd->layer.up, pkt); |
262 | ||
263 | /* For -EILSEQ the packet is not freed so so it now */ | |
264 | if (err == -EILSEQ) | |
265 | cfpkt_destroy(pkt); | |
bd30ce4b | 266 | |
267 | /* Release reference to stack upwards */ | |
268 | caifd_put(caifd); | |
7c18d220 | 269 | |
270 | if (err != 0) | |
271 | err = NET_RX_DROP; | |
272 | return err; | |
c72dfae2 SB |
273 | } |
274 | ||
275 | static struct packet_type caif_packet_type __read_mostly = { | |
276 | .type = cpu_to_be16(ETH_P_CAIF), | |
277 | .func = receive, | |
278 | }; | |
279 | ||
280 | static void dev_flowctrl(struct net_device *dev, int on) | |
281 | { | |
bd30ce4b | 282 | struct caif_device_entry *caifd; |
283 | ||
284 | rcu_read_lock(); | |
285 | ||
286 | caifd = caif_get(dev); | |
287 | if (!caifd || !caifd->layer.up || !caifd->layer.up->ctrlcmd) { | |
288 | rcu_read_unlock(); | |
c72dfae2 | 289 | return; |
bd30ce4b | 290 | } |
291 | ||
292 | caifd_hold(caifd); | |
293 | rcu_read_unlock(); | |
c72dfae2 SB |
294 | |
295 | caifd->layer.up->ctrlcmd(caifd->layer.up, | |
296 | on ? | |
297 | _CAIF_CTRLCMD_PHYIF_FLOW_ON_IND : | |
298 | _CAIF_CTRLCMD_PHYIF_FLOW_OFF_IND, | |
299 | caifd->layer.id); | |
bd30ce4b | 300 | caifd_put(caifd); |
c72dfae2 SB |
301 | } |
302 | ||
7c18d220 | 303 | void caif_enroll_dev(struct net_device *dev, struct caif_dev_common *caifdev, |
304 | struct cflayer *link_support, int head_room, | |
305 | struct cflayer **layer, int (**rcv_func)( | |
306 | struct sk_buff *, struct net_device *, | |
307 | struct packet_type *, struct net_device *)) | |
308 | { | |
309 | struct caif_device_entry *caifd; | |
310 | enum cfcnfg_phy_preference pref; | |
311 | struct cfcnfg *cfg = get_cfcnfg(dev_net(dev)); | |
312 | struct caif_device_entry_list *caifdevs; | |
313 | ||
314 | caifdevs = caif_device_list(dev_net(dev)); | |
7c18d220 | 315 | caifd = caif_device_alloc(dev); |
316 | if (!caifd) | |
317 | return; | |
318 | *layer = &caifd->layer; | |
0e4c7d85 | 319 | spin_lock_init(&caifd->flow_lock); |
7c18d220 | 320 | |
321 | switch (caifdev->link_select) { | |
322 | case CAIF_LINK_HIGH_BANDW: | |
323 | pref = CFPHYPREF_HIGH_BW; | |
324 | break; | |
325 | case CAIF_LINK_LOW_LATENCY: | |
326 | pref = CFPHYPREF_LOW_LAT; | |
327 | break; | |
328 | default: | |
329 | pref = CFPHYPREF_HIGH_BW; | |
330 | break; | |
331 | } | |
332 | mutex_lock(&caifdevs->lock); | |
333 | list_add_rcu(&caifd->list, &caifdevs->list); | |
334 | ||
335 | strncpy(caifd->layer.name, dev->name, | |
336 | sizeof(caifd->layer.name) - 1); | |
337 | caifd->layer.name[sizeof(caifd->layer.name) - 1] = 0; | |
338 | caifd->layer.transmit = transmit; | |
339 | cfcnfg_add_phy_layer(cfg, | |
340 | dev, | |
341 | &caifd->layer, | |
342 | pref, | |
343 | link_support, | |
344 | caifdev->use_fcs, | |
345 | head_room); | |
346 | mutex_unlock(&caifdevs->lock); | |
347 | if (rcv_func) | |
348 | *rcv_func = receive; | |
349 | } | |
7ad65bf6 | 350 | EXPORT_SYMBOL(caif_enroll_dev); |
7c18d220 | 351 | |
c72dfae2 SB |
352 | /* notify Caif of device events */ |
353 | static int caif_device_notify(struct notifier_block *me, unsigned long what, | |
354 | void *arg) | |
355 | { | |
356 | struct net_device *dev = arg; | |
357 | struct caif_device_entry *caifd = NULL; | |
358 | struct caif_dev_common *caifdev; | |
bee925db | 359 | struct cfcnfg *cfg; |
7c18d220 | 360 | struct cflayer *layer, *link_support; |
361 | int head_room = 0; | |
08613e46 | 362 | struct caif_device_entry_list *caifdevs; |
c72dfae2 | 363 | |
bee925db | 364 | cfg = get_cfcnfg(dev_net(dev)); |
7c18d220 | 365 | caifdevs = caif_device_list(dev_net(dev)); |
bee925db | 366 | |
7c18d220 | 367 | caifd = caif_get(dev); |
368 | if (caifd == NULL && dev->type != ARPHRD_CAIF) | |
369 | return 0; | |
08613e46 | 370 | |
c72dfae2 SB |
371 | switch (what) { |
372 | case NETDEV_REGISTER: | |
7c18d220 | 373 | if (caifd != NULL) |
374 | break; | |
bd30ce4b | 375 | |
c72dfae2 | 376 | caifdev = netdev_priv(dev); |
c72dfae2 | 377 | |
7c18d220 | 378 | link_support = NULL; |
379 | if (caifdev->use_frag) { | |
380 | head_room = 1; | |
381 | link_support = cfserl_create(dev->ifindex, | |
e977b4cf | 382 | caifdev->use_stx); |
7c18d220 | 383 | if (!link_support) { |
384 | pr_warn("Out of memory\n"); | |
385 | break; | |
386 | } | |
c72dfae2 | 387 | } |
7c18d220 | 388 | caif_enroll_dev(dev, caifdev, link_support, head_room, |
389 | &layer, NULL); | |
390 | caifdev->flowctrl = dev_flowctrl; | |
c72dfae2 SB |
391 | break; |
392 | ||
bd30ce4b | 393 | case NETDEV_UP: |
394 | rcu_read_lock(); | |
395 | ||
c72dfae2 | 396 | caifd = caif_get(dev); |
bd30ce4b | 397 | if (caifd == NULL) { |
398 | rcu_read_unlock(); | |
c72dfae2 | 399 | break; |
bd30ce4b | 400 | } |
c72dfae2 | 401 | |
0e4c7d85 | 402 | caifd->xoff = 0; |
bd30ce4b | 403 | cfcnfg_set_phy_state(cfg, &caifd->layer, true); |
404 | rcu_read_unlock(); | |
c72dfae2 | 405 | |
c72dfae2 SB |
406 | break; |
407 | ||
408 | case NETDEV_DOWN: | |
bd30ce4b | 409 | rcu_read_lock(); |
410 | ||
c72dfae2 | 411 | caifd = caif_get(dev); |
bd30ce4b | 412 | if (!caifd || !caifd->layer.up || !caifd->layer.up->ctrlcmd) { |
413 | rcu_read_unlock(); | |
414 | return -EINVAL; | |
415 | } | |
416 | ||
417 | cfcnfg_set_phy_state(cfg, &caifd->layer, false); | |
418 | caifd_hold(caifd); | |
419 | rcu_read_unlock(); | |
420 | ||
421 | caifd->layer.up->ctrlcmd(caifd->layer.up, | |
422 | _CAIF_CTRLCMD_PHYIF_DOWN_IND, | |
423 | caifd->layer.id); | |
7d311304 | 424 | |
425 | spin_lock_bh(&caifd->flow_lock); | |
426 | ||
427 | /* | |
428 | * Replace our xoff-destructor with original destructor. | |
429 | * We trust that skb->destructor *always* is called before | |
430 | * the skb reference is invalid. The hijacked SKB destructor | |
431 | * takes the flow_lock so manipulating the skb->destructor here | |
432 | * should be safe. | |
433 | */ | |
434 | if (caifd->xoff_skb_dtor != NULL && caifd->xoff_skb != NULL) | |
435 | caifd->xoff_skb->destructor = caifd->xoff_skb_dtor; | |
436 | ||
437 | caifd->xoff = 0; | |
438 | caifd->xoff_skb_dtor = NULL; | |
439 | caifd->xoff_skb = NULL; | |
440 | ||
441 | spin_unlock_bh(&caifd->flow_lock); | |
bd30ce4b | 442 | caifd_put(caifd); |
c72dfae2 SB |
443 | break; |
444 | ||
445 | case NETDEV_UNREGISTER: | |
bd30ce4b | 446 | mutex_lock(&caifdevs->lock); |
447 | ||
c72dfae2 | 448 | caifd = caif_get(dev); |
bd30ce4b | 449 | if (caifd == NULL) { |
450 | mutex_unlock(&caifdevs->lock); | |
f2527ec4 | 451 | break; |
bd30ce4b | 452 | } |
453 | list_del_rcu(&caifd->list); | |
454 | ||
455 | /* | |
456 | * NETDEV_UNREGISTER is called repeatedly until all reference | |
457 | * counts for the net-device are released. If references to | |
458 | * caifd is taken, simply ignore NETDEV_UNREGISTER and wait for | |
459 | * the next call to NETDEV_UNREGISTER. | |
460 | * | |
461 | * If any packets are in flight down the CAIF Stack, | |
462 | * cfcnfg_del_phy_layer will return nonzero. | |
463 | * If no packets are in flight, the CAIF Stack associated | |
464 | * with the net-device un-registering is freed. | |
465 | */ | |
466 | ||
467 | if (caifd_refcnt_read(caifd) != 0 || | |
468 | cfcnfg_del_phy_layer(cfg, &caifd->layer) != 0) { | |
469 | ||
470 | pr_info("Wait for device inuse\n"); | |
471 | /* Enrole device if CAIF Stack is still in use */ | |
472 | list_add_rcu(&caifd->list, &caifdevs->list); | |
473 | mutex_unlock(&caifdevs->lock); | |
474 | break; | |
475 | } | |
476 | ||
477 | synchronize_rcu(); | |
478 | dev_put(caifd->netdev); | |
479 | free_percpu(caifd->pcpu_refcnt); | |
480 | kfree(caifd); | |
481 | ||
482 | mutex_unlock(&caifdevs->lock); | |
c72dfae2 SB |
483 | break; |
484 | } | |
485 | return 0; | |
486 | } | |
487 | ||
488 | static struct notifier_block caif_device_notifier = { | |
489 | .notifier_call = caif_device_notify, | |
490 | .priority = 0, | |
491 | }; | |
492 | ||
c72dfae2 SB |
493 | /* Per-namespace Caif devices handling */ |
494 | static int caif_init_net(struct net *net) | |
495 | { | |
496 | struct caif_net *caifn = net_generic(net, caif_net_id); | |
497 | INIT_LIST_HEAD(&caifn->caifdevs.list); | |
bd30ce4b | 498 | mutex_init(&caifn->caifdevs.lock); |
bee925db | 499 | |
500 | caifn->cfg = cfcnfg_create(); | |
f84ea779 | 501 | if (!caifn->cfg) |
bee925db | 502 | return -ENOMEM; |
bee925db | 503 | |
c72dfae2 SB |
504 | return 0; |
505 | } | |
506 | ||
507 | static void caif_exit_net(struct net *net) | |
508 | { | |
bd30ce4b | 509 | struct caif_device_entry *caifd, *tmp; |
510 | struct caif_device_entry_list *caifdevs = | |
511 | caif_device_list(net); | |
7c18d220 | 512 | struct cfcnfg *cfg = get_cfcnfg(net); |
513 | ||
c72dfae2 | 514 | rtnl_lock(); |
bd30ce4b | 515 | mutex_lock(&caifdevs->lock); |
516 | ||
517 | list_for_each_entry_safe(caifd, tmp, &caifdevs->list, list) { | |
518 | int i = 0; | |
519 | list_del_rcu(&caifd->list); | |
520 | cfcnfg_set_phy_state(cfg, &caifd->layer, false); | |
521 | ||
522 | while (i < 10 && | |
523 | (caifd_refcnt_read(caifd) != 0 || | |
524 | cfcnfg_del_phy_layer(cfg, &caifd->layer) != 0)) { | |
525 | ||
526 | pr_info("Wait for device inuse\n"); | |
527 | msleep(250); | |
528 | i++; | |
529 | } | |
530 | synchronize_rcu(); | |
531 | dev_put(caifd->netdev); | |
532 | free_percpu(caifd->pcpu_refcnt); | |
533 | kfree(caifd); | |
c72dfae2 | 534 | } |
bee925db | 535 | cfcnfg_remove(cfg); |
bd30ce4b | 536 | |
537 | mutex_unlock(&caifdevs->lock); | |
c72dfae2 SB |
538 | rtnl_unlock(); |
539 | } | |
540 | ||
541 | static struct pernet_operations caif_net_ops = { | |
542 | .init = caif_init_net, | |
543 | .exit = caif_exit_net, | |
544 | .id = &caif_net_id, | |
545 | .size = sizeof(struct caif_net), | |
546 | }; | |
547 | ||
548 | /* Initialize Caif devices list */ | |
549 | static int __init caif_device_init(void) | |
550 | { | |
551 | int result; | |
bd30ce4b | 552 | |
8a8ee9af | 553 | result = register_pernet_subsys(&caif_net_ops); |
c72dfae2 | 554 | |
bee925db | 555 | if (result) |
c72dfae2 | 556 | return result; |
bee925db | 557 | |
c72dfae2 | 558 | register_netdevice_notifier(&caif_device_notifier); |
bee925db | 559 | dev_add_pack(&caif_packet_type); |
c72dfae2 SB |
560 | |
561 | return result; | |
c72dfae2 SB |
562 | } |
563 | ||
564 | static void __exit caif_device_exit(void) | |
565 | { | |
c72dfae2 | 566 | unregister_netdevice_notifier(&caif_device_notifier); |
bee925db | 567 | dev_remove_pack(&caif_packet_type); |
96f80d12 | 568 | unregister_pernet_subsys(&caif_net_ops); |
c72dfae2 SB |
569 | } |
570 | ||
571 | module_init(caif_device_init); | |
572 | module_exit(caif_device_exit); |