[NET] X25: Fix whitespace errors.
[linux-2.6-block.git] / net / xfrm / xfrm_algo.c
CommitLineData
1da177e4
LT
1/*
2 * xfrm algorithm interface
3 *
4 * Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
5 *
6 * This program is free software; you can redistribute it and/or modify it
7 * under the terms of the GNU General Public License as published by the Free
8 * Software Foundation; either version 2 of the License, or (at your option)
9 * any later version.
10 */
11
1da177e4
LT
12#include <linux/module.h>
13#include <linux/kernel.h>
14#include <linux/pfkeyv2.h>
15#include <linux/crypto.h>
16#include <net/xfrm.h>
17#if defined(CONFIG_INET_AH) || defined(CONFIG_INET_AH_MODULE) || defined(CONFIG_INET6_AH) || defined(CONFIG_INET6_AH_MODULE)
18#include <net/ah.h>
19#endif
20#if defined(CONFIG_INET_ESP) || defined(CONFIG_INET_ESP_MODULE) || defined(CONFIG_INET6_ESP) || defined(CONFIG_INET6_ESP_MODULE)
21#include <net/esp.h>
22#endif
23#include <asm/scatterlist.h>
24
25/*
26 * Algorithms supported by IPsec. These entries contain properties which
27 * are used in key negotiation and xfrm processing, and are used to verify
28 * that instantiated crypto transforms have correct parameters for IPsec
29 * purposes.
30 */
31static struct xfrm_algo_desc aalg_list[] = {
32{
07d4ee58
HX
33 .name = "hmac(digest_null)",
34 .compat = "digest_null",
1da177e4
LT
35
36 .uinfo = {
37 .auth = {
38 .icv_truncbits = 0,
39 .icv_fullbits = 0,
40 }
41 },
42
43 .desc = {
44 .sadb_alg_id = SADB_X_AALG_NULL,
45 .sadb_alg_ivlen = 0,
46 .sadb_alg_minbits = 0,
47 .sadb_alg_maxbits = 0
48 }
49},
50{
07d4ee58
HX
51 .name = "hmac(md5)",
52 .compat = "md5",
1da177e4
LT
53
54 .uinfo = {
55 .auth = {
56 .icv_truncbits = 96,
57 .icv_fullbits = 128,
58 }
59 },
60
61 .desc = {
62 .sadb_alg_id = SADB_AALG_MD5HMAC,
63 .sadb_alg_ivlen = 0,
64 .sadb_alg_minbits = 128,
65 .sadb_alg_maxbits = 128
66 }
67},
68{
07d4ee58
HX
69 .name = "hmac(sha1)",
70 .compat = "sha1",
1da177e4
LT
71
72 .uinfo = {
73 .auth = {
74 .icv_truncbits = 96,
75 .icv_fullbits = 160,
76 }
77 },
78
79 .desc = {
80 .sadb_alg_id = SADB_AALG_SHA1HMAC,
81 .sadb_alg_ivlen = 0,
82 .sadb_alg_minbits = 160,
83 .sadb_alg_maxbits = 160
84 }
85},
86{
07d4ee58
HX
87 .name = "hmac(sha256)",
88 .compat = "sha256",
1da177e4
LT
89
90 .uinfo = {
91 .auth = {
92 .icv_truncbits = 96,
93 .icv_fullbits = 256,
94 }
95 },
96
97 .desc = {
98 .sadb_alg_id = SADB_X_AALG_SHA2_256HMAC,
99 .sadb_alg_ivlen = 0,
100 .sadb_alg_minbits = 256,
101 .sadb_alg_maxbits = 256
102 }
103},
104{
07d4ee58
HX
105 .name = "hmac(ripemd160)",
106 .compat = "ripemd160",
1da177e4
LT
107
108 .uinfo = {
109 .auth = {
110 .icv_truncbits = 96,
111 .icv_fullbits = 160,
112 }
113 },
114
115 .desc = {
116 .sadb_alg_id = SADB_X_AALG_RIPEMD160HMAC,
117 .sadb_alg_ivlen = 0,
118 .sadb_alg_minbits = 160,
119 .sadb_alg_maxbits = 160
120 }
121},
7cf4c1a5
KM
122{
123 .name = "xcbc(aes)",
124
125 .uinfo = {
126 .auth = {
127 .icv_truncbits = 96,
128 .icv_fullbits = 128,
129 }
130 },
131
132 .desc = {
133 .sadb_alg_id = SADB_X_AALG_AES_XCBC_MAC,
134 .sadb_alg_ivlen = 0,
135 .sadb_alg_minbits = 128,
136 .sadb_alg_maxbits = 128
137 }
138},
1da177e4
LT
139};
140
141static struct xfrm_algo_desc ealg_list[] = {
142{
6b7326c8
HX
143 .name = "ecb(cipher_null)",
144 .compat = "cipher_null",
1da177e4
LT
145
146 .uinfo = {
147 .encr = {
148 .blockbits = 8,
149 .defkeybits = 0,
150 }
151 },
152
153 .desc = {
154 .sadb_alg_id = SADB_EALG_NULL,
155 .sadb_alg_ivlen = 0,
156 .sadb_alg_minbits = 0,
157 .sadb_alg_maxbits = 0
158 }
159},
160{
6b7326c8
HX
161 .name = "cbc(des)",
162 .compat = "des",
1da177e4
LT
163
164 .uinfo = {
165 .encr = {
166 .blockbits = 64,
167 .defkeybits = 64,
168 }
169 },
170
171 .desc = {
172 .sadb_alg_id = SADB_EALG_DESCBC,
173 .sadb_alg_ivlen = 8,
174 .sadb_alg_minbits = 64,
175 .sadb_alg_maxbits = 64
176 }
177},
178{
6b7326c8
HX
179 .name = "cbc(des3_ede)",
180 .compat = "des3_ede",
1da177e4
LT
181
182 .uinfo = {
183 .encr = {
184 .blockbits = 64,
185 .defkeybits = 192,
186 }
187 },
188
189 .desc = {
190 .sadb_alg_id = SADB_EALG_3DESCBC,
191 .sadb_alg_ivlen = 8,
192 .sadb_alg_minbits = 192,
193 .sadb_alg_maxbits = 192
194 }
195},
196{
6b7326c8
HX
197 .name = "cbc(cast128)",
198 .compat = "cast128",
1da177e4
LT
199
200 .uinfo = {
201 .encr = {
202 .blockbits = 64,
203 .defkeybits = 128,
204 }
205 },
206
207 .desc = {
208 .sadb_alg_id = SADB_X_EALG_CASTCBC,
209 .sadb_alg_ivlen = 8,
210 .sadb_alg_minbits = 40,
211 .sadb_alg_maxbits = 128
212 }
213},
214{
6b7326c8
HX
215 .name = "cbc(blowfish)",
216 .compat = "blowfish",
1da177e4
LT
217
218 .uinfo = {
219 .encr = {
220 .blockbits = 64,
221 .defkeybits = 128,
222 }
223 },
224
225 .desc = {
226 .sadb_alg_id = SADB_X_EALG_BLOWFISHCBC,
227 .sadb_alg_ivlen = 8,
228 .sadb_alg_minbits = 40,
229 .sadb_alg_maxbits = 448
230 }
231},
232{
6b7326c8
HX
233 .name = "cbc(aes)",
234 .compat = "aes",
1da177e4
LT
235
236 .uinfo = {
237 .encr = {
238 .blockbits = 128,
239 .defkeybits = 128,
240 }
241 },
242
243 .desc = {
244 .sadb_alg_id = SADB_X_EALG_AESCBC,
245 .sadb_alg_ivlen = 8,
246 .sadb_alg_minbits = 128,
247 .sadb_alg_maxbits = 256
248 }
249},
250{
6b7326c8
HX
251 .name = "cbc(serpent)",
252 .compat = "serpent",
1da177e4
LT
253
254 .uinfo = {
255 .encr = {
256 .blockbits = 128,
257 .defkeybits = 128,
258 }
259 },
260
261 .desc = {
262 .sadb_alg_id = SADB_X_EALG_SERPENTCBC,
263 .sadb_alg_ivlen = 8,
264 .sadb_alg_minbits = 128,
265 .sadb_alg_maxbits = 256,
266 }
267},
6a0dc8d7
NT
268{
269 .name = "cbc(camellia)",
270
271 .uinfo = {
272 .encr = {
273 .blockbits = 128,
274 .defkeybits = 128,
275 }
276 },
277
278 .desc = {
279 .sadb_alg_id = SADB_X_EALG_CAMELLIACBC,
280 .sadb_alg_ivlen = 8,
281 .sadb_alg_minbits = 128,
282 .sadb_alg_maxbits = 256
283 }
284},
1da177e4 285{
6b7326c8
HX
286 .name = "cbc(twofish)",
287 .compat = "twofish",
1da177e4
LT
288
289 .uinfo = {
290 .encr = {
291 .blockbits = 128,
292 .defkeybits = 128,
293 }
294 },
295
296 .desc = {
297 .sadb_alg_id = SADB_X_EALG_TWOFISHCBC,
298 .sadb_alg_ivlen = 8,
299 .sadb_alg_minbits = 128,
300 .sadb_alg_maxbits = 256
301 }
302},
303};
304
305static struct xfrm_algo_desc calg_list[] = {
306{
307 .name = "deflate",
308 .uinfo = {
309 .comp = {
310 .threshold = 90,
311 }
312 },
313 .desc = { .sadb_alg_id = SADB_X_CALG_DEFLATE }
314},
315{
316 .name = "lzs",
317 .uinfo = {
318 .comp = {
319 .threshold = 90,
320 }
321 },
322 .desc = { .sadb_alg_id = SADB_X_CALG_LZS }
323},
324{
325 .name = "lzjh",
326 .uinfo = {
327 .comp = {
328 .threshold = 50,
329 }
330 },
331 .desc = { .sadb_alg_id = SADB_X_CALG_LZJH }
332},
333};
334
335static inline int aalg_entries(void)
336{
337 return ARRAY_SIZE(aalg_list);
338}
339
340static inline int ealg_entries(void)
341{
342 return ARRAY_SIZE(ealg_list);
343}
344
345static inline int calg_entries(void)
346{
347 return ARRAY_SIZE(calg_list);
348}
349
350/* Todo: generic iterators */
351struct xfrm_algo_desc *xfrm_aalg_get_byid(int alg_id)
352{
353 int i;
354
355 for (i = 0; i < aalg_entries(); i++) {
356 if (aalg_list[i].desc.sadb_alg_id == alg_id) {
357 if (aalg_list[i].available)
358 return &aalg_list[i];
359 else
360 break;
361 }
362 }
363 return NULL;
364}
365EXPORT_SYMBOL_GPL(xfrm_aalg_get_byid);
366
367struct xfrm_algo_desc *xfrm_ealg_get_byid(int alg_id)
368{
369 int i;
370
371 for (i = 0; i < ealg_entries(); i++) {
372 if (ealg_list[i].desc.sadb_alg_id == alg_id) {
373 if (ealg_list[i].available)
374 return &ealg_list[i];
375 else
376 break;
377 }
378 }
379 return NULL;
380}
381EXPORT_SYMBOL_GPL(xfrm_ealg_get_byid);
382
383struct xfrm_algo_desc *xfrm_calg_get_byid(int alg_id)
384{
385 int i;
386
387 for (i = 0; i < calg_entries(); i++) {
388 if (calg_list[i].desc.sadb_alg_id == alg_id) {
389 if (calg_list[i].available)
390 return &calg_list[i];
391 else
392 break;
393 }
394 }
395 return NULL;
396}
397EXPORT_SYMBOL_GPL(xfrm_calg_get_byid);
398
399static struct xfrm_algo_desc *xfrm_get_byname(struct xfrm_algo_desc *list,
e4d5b79c
HX
400 int entries, u32 type, u32 mask,
401 char *name, int probe)
1da177e4
LT
402{
403 int i, status;
404
405 if (!name)
406 return NULL;
407
408 for (i = 0; i < entries; i++) {
04ff1260
HX
409 if (strcmp(name, list[i].name) &&
410 (!list[i].compat || strcmp(name, list[i].compat)))
1da177e4
LT
411 continue;
412
413 if (list[i].available)
414 return &list[i];
415
416 if (!probe)
417 break;
418
b836267a
MW
419 status = crypto_has_alg(list[i].name, type,
420 mask | CRYPTO_ALG_ASYNC);
1da177e4
LT
421 if (!status)
422 break;
423
424 list[i].available = status;
425 return &list[i];
426 }
427 return NULL;
428}
429
430struct xfrm_algo_desc *xfrm_aalg_get_byname(char *name, int probe)
431{
e4d5b79c
HX
432 return xfrm_get_byname(aalg_list, aalg_entries(),
433 CRYPTO_ALG_TYPE_HASH, CRYPTO_ALG_TYPE_HASH_MASK,
434 name, probe);
1da177e4
LT
435}
436EXPORT_SYMBOL_GPL(xfrm_aalg_get_byname);
437
438struct xfrm_algo_desc *xfrm_ealg_get_byname(char *name, int probe)
439{
e4d5b79c
HX
440 return xfrm_get_byname(ealg_list, ealg_entries(),
441 CRYPTO_ALG_TYPE_BLKCIPHER, CRYPTO_ALG_TYPE_MASK,
442 name, probe);
1da177e4
LT
443}
444EXPORT_SYMBOL_GPL(xfrm_ealg_get_byname);
445
446struct xfrm_algo_desc *xfrm_calg_get_byname(char *name, int probe)
447{
e4d5b79c
HX
448 return xfrm_get_byname(calg_list, calg_entries(),
449 CRYPTO_ALG_TYPE_COMPRESS, CRYPTO_ALG_TYPE_MASK,
450 name, probe);
1da177e4
LT
451}
452EXPORT_SYMBOL_GPL(xfrm_calg_get_byname);
453
454struct xfrm_algo_desc *xfrm_aalg_get_byidx(unsigned int idx)
455{
456 if (idx >= aalg_entries())
457 return NULL;
458
459 return &aalg_list[idx];
460}
461EXPORT_SYMBOL_GPL(xfrm_aalg_get_byidx);
462
463struct xfrm_algo_desc *xfrm_ealg_get_byidx(unsigned int idx)
464{
465 if (idx >= ealg_entries())
466 return NULL;
467
468 return &ealg_list[idx];
469}
470EXPORT_SYMBOL_GPL(xfrm_ealg_get_byidx);
471
472/*
473 * Probe for the availability of crypto algorithms, and set the available
474 * flag for any algorithms found on the system. This is typically called by
475 * pfkey during userspace SA add, update or register.
476 */
477void xfrm_probe_algs(void)
478{
479#ifdef CONFIG_CRYPTO
480 int i, status;
481
482 BUG_ON(in_softirq());
483
484 for (i = 0; i < aalg_entries(); i++) {
e4d5b79c
HX
485 status = crypto_has_hash(aalg_list[i].name, 0,
486 CRYPTO_ALG_ASYNC);
1da177e4
LT
487 if (aalg_list[i].available != status)
488 aalg_list[i].available = status;
489 }
490
491 for (i = 0; i < ealg_entries(); i++) {
e4d5b79c
HX
492 status = crypto_has_blkcipher(ealg_list[i].name, 0,
493 CRYPTO_ALG_ASYNC);
1da177e4
LT
494 if (ealg_list[i].available != status)
495 ealg_list[i].available = status;
496 }
497
498 for (i = 0; i < calg_entries(); i++) {
e4d5b79c
HX
499 status = crypto_has_comp(calg_list[i].name, 0,
500 CRYPTO_ALG_ASYNC);
1da177e4
LT
501 if (calg_list[i].available != status)
502 calg_list[i].available = status;
503 }
504#endif
505}
506EXPORT_SYMBOL_GPL(xfrm_probe_algs);
507
508int xfrm_count_auth_supported(void)
509{
510 int i, n;
511
512 for (i = 0, n = 0; i < aalg_entries(); i++)
513 if (aalg_list[i].available)
514 n++;
515 return n;
516}
517EXPORT_SYMBOL_GPL(xfrm_count_auth_supported);
518
519int xfrm_count_enc_supported(void)
520{
521 int i, n;
522
523 for (i = 0, n = 0; i < ealg_entries(); i++)
524 if (ealg_list[i].available)
525 n++;
526 return n;
527}
528EXPORT_SYMBOL_GPL(xfrm_count_enc_supported);
529
530/* Move to common area: it is shared with AH. */
531
07d4ee58
HX
532int skb_icv_walk(const struct sk_buff *skb, struct hash_desc *desc,
533 int offset, int len, icv_update_fn_t icv_update)
1da177e4
LT
534{
535 int start = skb_headlen(skb);
536 int i, copy = start - offset;
07d4ee58 537 int err;
1da177e4
LT
538 struct scatterlist sg;
539
540 /* Checksum header. */
541 if (copy > 0) {
542 if (copy > len)
543 copy = len;
544
545 sg.page = virt_to_page(skb->data + offset);
546 sg.offset = (unsigned long)(skb->data + offset) % PAGE_SIZE;
547 sg.length = copy;
548
07d4ee58
HX
549 err = icv_update(desc, &sg, copy);
550 if (unlikely(err))
551 return err;
1da177e4
LT
552
553 if ((len -= copy) == 0)
07d4ee58 554 return 0;
1da177e4
LT
555 offset += copy;
556 }
557
558 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
559 int end;
560
561 BUG_TRAP(start <= offset + len);
562
563 end = start + skb_shinfo(skb)->frags[i].size;
564 if ((copy = end - offset) > 0) {
565 skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
566
567 if (copy > len)
568 copy = len;
569
570 sg.page = frag->page;
571 sg.offset = frag->page_offset + offset-start;
572 sg.length = copy;
573
07d4ee58
HX
574 err = icv_update(desc, &sg, copy);
575 if (unlikely(err))
576 return err;
1da177e4
LT
577
578 if (!(len -= copy))
07d4ee58 579 return 0;
1da177e4
LT
580 offset += copy;
581 }
582 start = end;
583 }
584
585 if (skb_shinfo(skb)->frag_list) {
586 struct sk_buff *list = skb_shinfo(skb)->frag_list;
587
588 for (; list; list = list->next) {
589 int end;
590
591 BUG_TRAP(start <= offset + len);
592
593 end = start + list->len;
594 if ((copy = end - offset) > 0) {
595 if (copy > len)
596 copy = len;
07d4ee58
HX
597 err = skb_icv_walk(list, desc, offset-start,
598 copy, icv_update);
599 if (unlikely(err))
600 return err;
1da177e4 601 if ((len -= copy) == 0)
07d4ee58 602 return 0;
1da177e4
LT
603 offset += copy;
604 }
605 start = end;
606 }
607 }
09a62660 608 BUG_ON(len);
07d4ee58 609 return 0;
1da177e4
LT
610}
611EXPORT_SYMBOL_GPL(skb_icv_walk);
612
613#if defined(CONFIG_INET_ESP) || defined(CONFIG_INET_ESP_MODULE) || defined(CONFIG_INET6_ESP) || defined(CONFIG_INET6_ESP_MODULE)
614
615/* Looking generic it is not used in another places. */
616
617int
618skb_to_sgvec(struct sk_buff *skb, struct scatterlist *sg, int offset, int len)
619{
620 int start = skb_headlen(skb);
621 int i, copy = start - offset;
622 int elt = 0;
623
624 if (copy > 0) {
625 if (copy > len)
626 copy = len;
627 sg[elt].page = virt_to_page(skb->data + offset);
628 sg[elt].offset = (unsigned long)(skb->data + offset) % PAGE_SIZE;
629 sg[elt].length = copy;
630 elt++;
631 if ((len -= copy) == 0)
632 return elt;
633 offset += copy;
634 }
635
636 for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) {
637 int end;
638
639 BUG_TRAP(start <= offset + len);
640
641 end = start + skb_shinfo(skb)->frags[i].size;
642 if ((copy = end - offset) > 0) {
643 skb_frag_t *frag = &skb_shinfo(skb)->frags[i];
644
645 if (copy > len)
646 copy = len;
647 sg[elt].page = frag->page;
648 sg[elt].offset = frag->page_offset+offset-start;
649 sg[elt].length = copy;
650 elt++;
651 if (!(len -= copy))
652 return elt;
653 offset += copy;
654 }
655 start = end;
656 }
657
658 if (skb_shinfo(skb)->frag_list) {
659 struct sk_buff *list = skb_shinfo(skb)->frag_list;
660
661 for (; list; list = list->next) {
662 int end;
663
664 BUG_TRAP(start <= offset + len);
665
666 end = start + list->len;
667 if ((copy = end - offset) > 0) {
668 if (copy > len)
669 copy = len;
670 elt += skb_to_sgvec(list, sg+elt, offset - start, copy);
671 if ((len -= copy) == 0)
672 return elt;
673 offset += copy;
674 }
675 start = end;
676 }
677 }
09a62660 678 BUG_ON(len);
1da177e4
LT
679 return elt;
680}
681EXPORT_SYMBOL_GPL(skb_to_sgvec);
682
683/* Check that skb data bits are writable. If they are not, copy data
684 * to newly created private area. If "tailbits" is given, make sure that
685 * tailbits bytes beyond current end of skb are writable.
686 *
687 * Returns amount of elements of scatterlist to load for subsequent
688 * transformations and pointer to writable trailer skb.
689 */
690
691int skb_cow_data(struct sk_buff *skb, int tailbits, struct sk_buff **trailer)
692{
693 int copyflag;
694 int elt;
695 struct sk_buff *skb1, **skb_p;
696
697 /* If skb is cloned or its head is paged, reallocate
698 * head pulling out all the pages (pages are considered not writable
699 * at the moment even if they are anonymous).
700 */
701 if ((skb_cloned(skb) || skb_shinfo(skb)->nr_frags) &&
702 __pskb_pull_tail(skb, skb_pagelen(skb)-skb_headlen(skb)) == NULL)
703 return -ENOMEM;
704
705 /* Easy case. Most of packets will go this way. */
706 if (!skb_shinfo(skb)->frag_list) {
707 /* A little of trouble, not enough of space for trailer.
708 * This should not happen, when stack is tuned to generate
709 * good frames. OK, on miss we reallocate and reserve even more
710 * space, 128 bytes is fair. */
711
712 if (skb_tailroom(skb) < tailbits &&
713 pskb_expand_head(skb, 0, tailbits-skb_tailroom(skb)+128, GFP_ATOMIC))
714 return -ENOMEM;
715
716 /* Voila! */
717 *trailer = skb;
718 return 1;
719 }
720
721 /* Misery. We are in troubles, going to mincer fragments... */
722
723 elt = 1;
724 skb_p = &skb_shinfo(skb)->frag_list;
725 copyflag = 0;
726
727 while ((skb1 = *skb_p) != NULL) {
728 int ntail = 0;
729
730 /* The fragment is partially pulled by someone,
731 * this can happen on input. Copy it and everything
732 * after it. */
733
734 if (skb_shared(skb1))
735 copyflag = 1;
736
737 /* If the skb is the last, worry about trailer. */
738
739 if (skb1->next == NULL && tailbits) {
740 if (skb_shinfo(skb1)->nr_frags ||
741 skb_shinfo(skb1)->frag_list ||
742 skb_tailroom(skb1) < tailbits)
743 ntail = tailbits + 128;
744 }
745
746 if (copyflag ||
747 skb_cloned(skb1) ||
748 ntail ||
749 skb_shinfo(skb1)->nr_frags ||
750 skb_shinfo(skb1)->frag_list) {
751 struct sk_buff *skb2;
752
753 /* Fuck, we are miserable poor guys... */
754 if (ntail == 0)
755 skb2 = skb_copy(skb1, GFP_ATOMIC);
756 else
757 skb2 = skb_copy_expand(skb1,
758 skb_headroom(skb1),
759 ntail,
760 GFP_ATOMIC);
761 if (unlikely(skb2 == NULL))
762 return -ENOMEM;
763
764 if (skb1->sk)
d4810200 765 skb_set_owner_w(skb2, skb1->sk);
1da177e4
LT
766
767 /* Looking around. Are we still alive?
768 * OK, link new skb, drop old one */
769
770 skb2->next = skb1->next;
771 *skb_p = skb2;
772 kfree_skb(skb1);
773 skb1 = skb2;
774 }
775 elt++;
776 *trailer = skb1;
777 skb_p = &skb1->next;
778 }
779
780 return elt;
781}
782EXPORT_SYMBOL_GPL(skb_cow_data);
783
784void *pskb_put(struct sk_buff *skb, struct sk_buff *tail, int len)
785{
786 if (tail != skb) {
787 skb->data_len += len;
788 skb->len += len;
789 }
790 return skb_put(tail, len);
791}
792EXPORT_SYMBOL_GPL(pskb_put);
793#endif