1 // SPDX-License-Identifier: GPL-2.0-or-later
3 * Asynchronous Compression operations
5 * Copyright (c) 2016, Intel Corporation
6 * Authors: Weigang Li <weigang.li@intel.com>
7 * Giovanni Cabiddu <giovanni.cabiddu@intel.com>
10 #include <crypto/internal/acompress.h>
11 #include <linux/cryptouser.h>
12 #include <linux/errno.h>
13 #include <linux/kernel.h>
14 #include <linux/module.h>
15 #include <linux/seq_file.h>
16 #include <linux/slab.h>
17 #include <linux/string.h>
18 #include <net/netlink.h>
24 static const struct crypto_type crypto_acomp_type;
26 static inline struct acomp_alg *__crypto_acomp_alg(struct crypto_alg *alg)
28 return container_of(alg, struct acomp_alg, calg.base);
31 static inline struct acomp_alg *crypto_acomp_alg(struct crypto_acomp *tfm)
33 return __crypto_acomp_alg(crypto_acomp_tfm(tfm)->__crt_alg);
36 static int __maybe_unused crypto_acomp_report(
37 struct sk_buff *skb, struct crypto_alg *alg)
39 struct crypto_report_acomp racomp;
41 memset(&racomp, 0, sizeof(racomp));
43 strscpy(racomp.type, "acomp", sizeof(racomp.type));
45 return nla_put(skb, CRYPTOCFGA_REPORT_ACOMP, sizeof(racomp), &racomp);
48 static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
51 static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg)
53 seq_puts(m, "type : acomp\n");
56 static void crypto_acomp_exit_tfm(struct crypto_tfm *tfm)
58 struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
59 struct acomp_alg *alg = crypto_acomp_alg(acomp);
64 static int crypto_acomp_init_tfm(struct crypto_tfm *tfm)
66 struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm);
67 struct acomp_alg *alg = crypto_acomp_alg(acomp);
69 if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
70 return crypto_init_scomp_ops_async(tfm);
72 acomp->compress = alg->compress;
73 acomp->decompress = alg->decompress;
74 acomp->dst_free = alg->dst_free;
75 acomp->reqsize = alg->reqsize;
78 acomp->base.exit = crypto_acomp_exit_tfm;
81 return alg->init(acomp);
86 static unsigned int crypto_acomp_extsize(struct crypto_alg *alg)
88 int extsize = crypto_alg_extsize(alg);
90 if (alg->cra_type != &crypto_acomp_type)
91 extsize += sizeof(struct crypto_scomp *);
96 static const struct crypto_type crypto_acomp_type = {
97 .extsize = crypto_acomp_extsize,
98 .init_tfm = crypto_acomp_init_tfm,
100 .show = crypto_acomp_show,
102 #if IS_ENABLED(CONFIG_CRYPTO_USER)
103 .report = crypto_acomp_report,
105 .maskclear = ~CRYPTO_ALG_TYPE_MASK,
106 .maskset = CRYPTO_ALG_TYPE_ACOMPRESS_MASK,
107 .type = CRYPTO_ALG_TYPE_ACOMPRESS,
108 .tfmsize = offsetof(struct crypto_acomp, base),
111 struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type,
114 return crypto_alloc_tfm(alg_name, &crypto_acomp_type, type, mask);
116 EXPORT_SYMBOL_GPL(crypto_alloc_acomp);
118 struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type,
121 return crypto_alloc_tfm_node(alg_name, &crypto_acomp_type, type, mask,
124 EXPORT_SYMBOL_GPL(crypto_alloc_acomp_node);
126 struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp)
128 struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
129 struct acomp_req *req;
131 req = __acomp_request_alloc(acomp);
132 if (req && (tfm->__crt_alg->cra_type != &crypto_acomp_type))
133 return crypto_acomp_scomp_alloc_ctx(req);
137 EXPORT_SYMBOL_GPL(acomp_request_alloc);
139 void acomp_request_free(struct acomp_req *req)
141 struct crypto_acomp *acomp = crypto_acomp_reqtfm(req);
142 struct crypto_tfm *tfm = crypto_acomp_tfm(acomp);
144 if (tfm->__crt_alg->cra_type != &crypto_acomp_type)
145 crypto_acomp_scomp_free_ctx(req);
147 if (req->flags & CRYPTO_ACOMP_ALLOC_OUTPUT) {
148 acomp->dst_free(req->dst);
152 __acomp_request_free(req);
154 EXPORT_SYMBOL_GPL(acomp_request_free);
156 void comp_prepare_alg(struct comp_alg_common *alg)
158 struct crypto_alg *base = &alg->base;
160 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK;
163 int crypto_register_acomp(struct acomp_alg *alg)
165 struct crypto_alg *base = &alg->calg.base;
167 comp_prepare_alg(&alg->calg);
169 base->cra_type = &crypto_acomp_type;
170 base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS;
172 return crypto_register_alg(base);
174 EXPORT_SYMBOL_GPL(crypto_register_acomp);
176 void crypto_unregister_acomp(struct acomp_alg *alg)
178 crypto_unregister_alg(&alg->base);
180 EXPORT_SYMBOL_GPL(crypto_unregister_acomp);
182 int crypto_register_acomps(struct acomp_alg *algs, int count)
186 for (i = 0; i < count; i++) {
187 ret = crypto_register_acomp(&algs[i]);
195 for (--i; i >= 0; --i)
196 crypto_unregister_acomp(&algs[i]);
200 EXPORT_SYMBOL_GPL(crypto_register_acomps);
202 void crypto_unregister_acomps(struct acomp_alg *algs, int count)
206 for (i = count - 1; i >= 0; --i)
207 crypto_unregister_acomp(&algs[i]);
209 EXPORT_SYMBOL_GPL(crypto_unregister_acomps);
211 MODULE_LICENSE("GPL");
212 MODULE_DESCRIPTION("Asynchronous compression type");