Commit | Line | Data |
---|---|---|
bf754ae8 JG |
1 | /* |
2 | * Cryptographic API. | |
3 | * | |
4 | * s390 implementation of the AES Cipher Algorithm. | |
5 | * | |
6 | * s390 Version: | |
86aa9fc2 | 7 | * Copyright IBM Corp. 2005,2007 |
bf754ae8 JG |
8 | * Author(s): Jan Glauber (jang@de.ibm.com) |
9 | * | |
10 | * Derived from "crypto/aes.c" | |
11 | * | |
12 | * This program is free software; you can redistribute it and/or modify it | |
13 | * under the terms of the GNU General Public License as published by the Free | |
14 | * Software Foundation; either version 2 of the License, or (at your option) | |
15 | * any later version. | |
16 | * | |
17 | */ | |
18 | ||
a9e62fad | 19 | #include <crypto/algapi.h> |
bf754ae8 JG |
20 | #include <linux/module.h> |
21 | #include <linux/init.h> | |
bf754ae8 JG |
22 | #include "crypt_s390.h" |
23 | ||
24 | #define AES_MIN_KEY_SIZE 16 | |
25 | #define AES_MAX_KEY_SIZE 32 | |
26 | ||
27 | /* data block size for all key lengths */ | |
28 | #define AES_BLOCK_SIZE 16 | |
29 | ||
86aa9fc2 JG |
30 | #define AES_KEYLEN_128 1 |
31 | #define AES_KEYLEN_192 2 | |
32 | #define AES_KEYLEN_256 4 | |
33 | ||
34 | static char keylen_flag = 0; | |
bf754ae8 JG |
35 | |
36 | struct s390_aes_ctx { | |
37 | u8 iv[AES_BLOCK_SIZE]; | |
38 | u8 key[AES_MAX_KEY_SIZE]; | |
a9e62fad HX |
39 | long enc; |
40 | long dec; | |
bf754ae8 JG |
41 | int key_len; |
42 | }; | |
43 | ||
6c2bb98b | 44 | static int aes_set_key(struct crypto_tfm *tfm, const u8 *in_key, |
560c06ae | 45 | unsigned int key_len) |
bf754ae8 | 46 | { |
6c2bb98b | 47 | struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); |
560c06ae | 48 | u32 *flags = &tfm->crt_flags; |
bf754ae8 JG |
49 | |
50 | switch (key_len) { | |
51 | case 16: | |
86aa9fc2 | 52 | if (!(keylen_flag & AES_KEYLEN_128)) |
bf754ae8 JG |
53 | goto fail; |
54 | break; | |
55 | case 24: | |
86aa9fc2 | 56 | if (!(keylen_flag & AES_KEYLEN_192)) |
bf754ae8 JG |
57 | goto fail; |
58 | ||
59 | break; | |
60 | case 32: | |
86aa9fc2 | 61 | if (!(keylen_flag & AES_KEYLEN_256)) |
bf754ae8 JG |
62 | goto fail; |
63 | break; | |
64 | default: | |
bf754ae8 JG |
65 | goto fail; |
66 | break; | |
67 | } | |
68 | ||
69 | sctx->key_len = key_len; | |
70 | memcpy(sctx->key, in_key, key_len); | |
71 | return 0; | |
72 | fail: | |
73 | *flags |= CRYPTO_TFM_RES_BAD_KEY_LEN; | |
74 | return -EINVAL; | |
75 | } | |
76 | ||
6c2bb98b | 77 | static void aes_encrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) |
bf754ae8 | 78 | { |
6c2bb98b | 79 | const struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); |
bf754ae8 JG |
80 | |
81 | switch (sctx->key_len) { | |
82 | case 16: | |
83 | crypt_s390_km(KM_AES_128_ENCRYPT, &sctx->key, out, in, | |
84 | AES_BLOCK_SIZE); | |
85 | break; | |
86 | case 24: | |
87 | crypt_s390_km(KM_AES_192_ENCRYPT, &sctx->key, out, in, | |
88 | AES_BLOCK_SIZE); | |
89 | break; | |
90 | case 32: | |
91 | crypt_s390_km(KM_AES_256_ENCRYPT, &sctx->key, out, in, | |
92 | AES_BLOCK_SIZE); | |
93 | break; | |
94 | } | |
95 | } | |
96 | ||
6c2bb98b | 97 | static void aes_decrypt(struct crypto_tfm *tfm, u8 *out, const u8 *in) |
bf754ae8 | 98 | { |
6c2bb98b | 99 | const struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); |
bf754ae8 JG |
100 | |
101 | switch (sctx->key_len) { | |
102 | case 16: | |
103 | crypt_s390_km(KM_AES_128_DECRYPT, &sctx->key, out, in, | |
104 | AES_BLOCK_SIZE); | |
105 | break; | |
106 | case 24: | |
107 | crypt_s390_km(KM_AES_192_DECRYPT, &sctx->key, out, in, | |
108 | AES_BLOCK_SIZE); | |
109 | break; | |
110 | case 32: | |
111 | crypt_s390_km(KM_AES_256_DECRYPT, &sctx->key, out, in, | |
112 | AES_BLOCK_SIZE); | |
113 | break; | |
114 | } | |
115 | } | |
116 | ||
bf754ae8 JG |
117 | |
118 | static struct crypto_alg aes_alg = { | |
119 | .cra_name = "aes", | |
65b75c36 HX |
120 | .cra_driver_name = "aes-s390", |
121 | .cra_priority = CRYPT_S390_PRIORITY, | |
f67d1369 JG |
122 | .cra_flags = CRYPTO_ALG_TYPE_CIPHER | |
123 | CRYPTO_ALG_NEED_FALLBACK, | |
bf754ae8 JG |
124 | .cra_blocksize = AES_BLOCK_SIZE, |
125 | .cra_ctxsize = sizeof(struct s390_aes_ctx), | |
126 | .cra_module = THIS_MODULE, | |
127 | .cra_list = LIST_HEAD_INIT(aes_alg.cra_list), | |
128 | .cra_u = { | |
129 | .cipher = { | |
130 | .cia_min_keysize = AES_MIN_KEY_SIZE, | |
131 | .cia_max_keysize = AES_MAX_KEY_SIZE, | |
132 | .cia_setkey = aes_set_key, | |
133 | .cia_encrypt = aes_encrypt, | |
134 | .cia_decrypt = aes_decrypt, | |
bf754ae8 JG |
135 | } |
136 | } | |
137 | }; | |
138 | ||
a9e62fad HX |
139 | static int ecb_aes_set_key(struct crypto_tfm *tfm, const u8 *in_key, |
140 | unsigned int key_len) | |
141 | { | |
142 | struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); | |
143 | ||
144 | switch (key_len) { | |
145 | case 16: | |
146 | sctx->enc = KM_AES_128_ENCRYPT; | |
147 | sctx->dec = KM_AES_128_DECRYPT; | |
148 | break; | |
149 | case 24: | |
150 | sctx->enc = KM_AES_192_ENCRYPT; | |
151 | sctx->dec = KM_AES_192_DECRYPT; | |
152 | break; | |
153 | case 32: | |
154 | sctx->enc = KM_AES_256_ENCRYPT; | |
155 | sctx->dec = KM_AES_256_DECRYPT; | |
156 | break; | |
157 | } | |
158 | ||
159 | return aes_set_key(tfm, in_key, key_len); | |
160 | } | |
161 | ||
162 | static int ecb_aes_crypt(struct blkcipher_desc *desc, long func, void *param, | |
163 | struct blkcipher_walk *walk) | |
164 | { | |
165 | int ret = blkcipher_walk_virt(desc, walk); | |
166 | unsigned int nbytes; | |
167 | ||
168 | while ((nbytes = walk->nbytes)) { | |
169 | /* only use complete blocks */ | |
170 | unsigned int n = nbytes & ~(AES_BLOCK_SIZE - 1); | |
171 | u8 *out = walk->dst.virt.addr; | |
172 | u8 *in = walk->src.virt.addr; | |
173 | ||
174 | ret = crypt_s390_km(func, param, out, in, n); | |
175 | BUG_ON((ret < 0) || (ret != n)); | |
176 | ||
177 | nbytes &= AES_BLOCK_SIZE - 1; | |
178 | ret = blkcipher_walk_done(desc, walk, nbytes); | |
179 | } | |
180 | ||
181 | return ret; | |
182 | } | |
183 | ||
184 | static int ecb_aes_encrypt(struct blkcipher_desc *desc, | |
185 | struct scatterlist *dst, struct scatterlist *src, | |
186 | unsigned int nbytes) | |
187 | { | |
188 | struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); | |
189 | struct blkcipher_walk walk; | |
190 | ||
191 | blkcipher_walk_init(&walk, dst, src, nbytes); | |
192 | return ecb_aes_crypt(desc, sctx->enc, sctx->key, &walk); | |
193 | } | |
194 | ||
195 | static int ecb_aes_decrypt(struct blkcipher_desc *desc, | |
196 | struct scatterlist *dst, struct scatterlist *src, | |
197 | unsigned int nbytes) | |
198 | { | |
199 | struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); | |
200 | struct blkcipher_walk walk; | |
201 | ||
202 | blkcipher_walk_init(&walk, dst, src, nbytes); | |
203 | return ecb_aes_crypt(desc, sctx->dec, sctx->key, &walk); | |
204 | } | |
205 | ||
206 | static struct crypto_alg ecb_aes_alg = { | |
207 | .cra_name = "ecb(aes)", | |
208 | .cra_driver_name = "ecb-aes-s390", | |
209 | .cra_priority = CRYPT_S390_COMPOSITE_PRIORITY, | |
f67d1369 JG |
210 | .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER | |
211 | CRYPTO_ALG_NEED_FALLBACK, | |
a9e62fad HX |
212 | .cra_blocksize = AES_BLOCK_SIZE, |
213 | .cra_ctxsize = sizeof(struct s390_aes_ctx), | |
214 | .cra_type = &crypto_blkcipher_type, | |
215 | .cra_module = THIS_MODULE, | |
216 | .cra_list = LIST_HEAD_INIT(ecb_aes_alg.cra_list), | |
217 | .cra_u = { | |
218 | .blkcipher = { | |
219 | .min_keysize = AES_MIN_KEY_SIZE, | |
220 | .max_keysize = AES_MAX_KEY_SIZE, | |
221 | .setkey = ecb_aes_set_key, | |
222 | .encrypt = ecb_aes_encrypt, | |
223 | .decrypt = ecb_aes_decrypt, | |
224 | } | |
225 | } | |
226 | }; | |
227 | ||
228 | static int cbc_aes_set_key(struct crypto_tfm *tfm, const u8 *in_key, | |
229 | unsigned int key_len) | |
230 | { | |
231 | struct s390_aes_ctx *sctx = crypto_tfm_ctx(tfm); | |
232 | ||
233 | switch (key_len) { | |
234 | case 16: | |
235 | sctx->enc = KMC_AES_128_ENCRYPT; | |
236 | sctx->dec = KMC_AES_128_DECRYPT; | |
237 | break; | |
238 | case 24: | |
239 | sctx->enc = KMC_AES_192_ENCRYPT; | |
240 | sctx->dec = KMC_AES_192_DECRYPT; | |
241 | break; | |
242 | case 32: | |
243 | sctx->enc = KMC_AES_256_ENCRYPT; | |
244 | sctx->dec = KMC_AES_256_DECRYPT; | |
245 | break; | |
246 | } | |
247 | ||
248 | return aes_set_key(tfm, in_key, key_len); | |
249 | } | |
250 | ||
251 | static int cbc_aes_crypt(struct blkcipher_desc *desc, long func, void *param, | |
252 | struct blkcipher_walk *walk) | |
253 | { | |
254 | int ret = blkcipher_walk_virt(desc, walk); | |
255 | unsigned int nbytes = walk->nbytes; | |
256 | ||
257 | if (!nbytes) | |
258 | goto out; | |
259 | ||
260 | memcpy(param, walk->iv, AES_BLOCK_SIZE); | |
261 | do { | |
262 | /* only use complete blocks */ | |
263 | unsigned int n = nbytes & ~(AES_BLOCK_SIZE - 1); | |
264 | u8 *out = walk->dst.virt.addr; | |
265 | u8 *in = walk->src.virt.addr; | |
266 | ||
267 | ret = crypt_s390_kmc(func, param, out, in, n); | |
268 | BUG_ON((ret < 0) || (ret != n)); | |
269 | ||
270 | nbytes &= AES_BLOCK_SIZE - 1; | |
271 | ret = blkcipher_walk_done(desc, walk, nbytes); | |
272 | } while ((nbytes = walk->nbytes)); | |
273 | memcpy(walk->iv, param, AES_BLOCK_SIZE); | |
274 | ||
275 | out: | |
276 | return ret; | |
277 | } | |
278 | ||
279 | static int cbc_aes_encrypt(struct blkcipher_desc *desc, | |
280 | struct scatterlist *dst, struct scatterlist *src, | |
281 | unsigned int nbytes) | |
282 | { | |
283 | struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); | |
284 | struct blkcipher_walk walk; | |
285 | ||
286 | blkcipher_walk_init(&walk, dst, src, nbytes); | |
287 | return cbc_aes_crypt(desc, sctx->enc, sctx->iv, &walk); | |
288 | } | |
289 | ||
290 | static int cbc_aes_decrypt(struct blkcipher_desc *desc, | |
291 | struct scatterlist *dst, struct scatterlist *src, | |
292 | unsigned int nbytes) | |
293 | { | |
294 | struct s390_aes_ctx *sctx = crypto_blkcipher_ctx(desc->tfm); | |
295 | struct blkcipher_walk walk; | |
296 | ||
297 | blkcipher_walk_init(&walk, dst, src, nbytes); | |
298 | return cbc_aes_crypt(desc, sctx->dec, sctx->iv, &walk); | |
299 | } | |
300 | ||
301 | static struct crypto_alg cbc_aes_alg = { | |
302 | .cra_name = "cbc(aes)", | |
303 | .cra_driver_name = "cbc-aes-s390", | |
304 | .cra_priority = CRYPT_S390_COMPOSITE_PRIORITY, | |
f67d1369 JG |
305 | .cra_flags = CRYPTO_ALG_TYPE_BLKCIPHER | |
306 | CRYPTO_ALG_NEED_FALLBACK, | |
a9e62fad HX |
307 | .cra_blocksize = AES_BLOCK_SIZE, |
308 | .cra_ctxsize = sizeof(struct s390_aes_ctx), | |
309 | .cra_type = &crypto_blkcipher_type, | |
310 | .cra_module = THIS_MODULE, | |
311 | .cra_list = LIST_HEAD_INIT(cbc_aes_alg.cra_list), | |
312 | .cra_u = { | |
313 | .blkcipher = { | |
314 | .min_keysize = AES_MIN_KEY_SIZE, | |
315 | .max_keysize = AES_MAX_KEY_SIZE, | |
316 | .ivsize = AES_BLOCK_SIZE, | |
317 | .setkey = cbc_aes_set_key, | |
318 | .encrypt = cbc_aes_encrypt, | |
319 | .decrypt = cbc_aes_decrypt, | |
320 | } | |
321 | } | |
322 | }; | |
323 | ||
bf754ae8 JG |
324 | static int __init aes_init(void) |
325 | { | |
326 | int ret; | |
327 | ||
328 | if (crypt_s390_func_available(KM_AES_128_ENCRYPT)) | |
86aa9fc2 | 329 | keylen_flag |= AES_KEYLEN_128; |
bf754ae8 | 330 | if (crypt_s390_func_available(KM_AES_192_ENCRYPT)) |
86aa9fc2 | 331 | keylen_flag |= AES_KEYLEN_192; |
bf754ae8 | 332 | if (crypt_s390_func_available(KM_AES_256_ENCRYPT)) |
86aa9fc2 JG |
333 | keylen_flag |= AES_KEYLEN_256; |
334 | ||
335 | if (!keylen_flag) | |
336 | return -EOPNOTSUPP; | |
bf754ae8 | 337 | |
86aa9fc2 | 338 | /* z9 109 and z9 BC/EC only support 128 bit key length */ |
f67d1369 JG |
339 | if (keylen_flag == AES_KEYLEN_128) { |
340 | aes_alg.cra_u.cipher.cia_max_keysize = AES_MIN_KEY_SIZE; | |
341 | ecb_aes_alg.cra_u.blkcipher.max_keysize = AES_MIN_KEY_SIZE; | |
342 | cbc_aes_alg.cra_u.blkcipher.max_keysize = AES_MIN_KEY_SIZE; | |
86aa9fc2 JG |
343 | printk(KERN_INFO |
344 | "aes_s390: hardware acceleration only available for" | |
345 | "128 bit keys\n"); | |
f67d1369 | 346 | } |
bf754ae8 JG |
347 | |
348 | ret = crypto_register_alg(&aes_alg); | |
86aa9fc2 | 349 | if (ret) |
a9e62fad | 350 | goto aes_err; |
a9e62fad HX |
351 | |
352 | ret = crypto_register_alg(&ecb_aes_alg); | |
86aa9fc2 | 353 | if (ret) |
a9e62fad | 354 | goto ecb_aes_err; |
a9e62fad HX |
355 | |
356 | ret = crypto_register_alg(&cbc_aes_alg); | |
86aa9fc2 | 357 | if (ret) |
a9e62fad | 358 | goto cbc_aes_err; |
a9e62fad HX |
359 | |
360 | out: | |
bf754ae8 | 361 | return ret; |
a9e62fad HX |
362 | |
363 | cbc_aes_err: | |
364 | crypto_unregister_alg(&ecb_aes_alg); | |
365 | ecb_aes_err: | |
366 | crypto_unregister_alg(&aes_alg); | |
367 | aes_err: | |
368 | goto out; | |
bf754ae8 JG |
369 | } |
370 | ||
371 | static void __exit aes_fini(void) | |
372 | { | |
a9e62fad HX |
373 | crypto_unregister_alg(&cbc_aes_alg); |
374 | crypto_unregister_alg(&ecb_aes_alg); | |
bf754ae8 JG |
375 | crypto_unregister_alg(&aes_alg); |
376 | } | |
377 | ||
378 | module_init(aes_init); | |
379 | module_exit(aes_fini); | |
380 | ||
381 | MODULE_ALIAS("aes"); | |
382 | ||
383 | MODULE_DESCRIPTION("Rijndael (AES) Cipher Algorithm"); | |
384 | MODULE_LICENSE("GPL"); | |
385 |