1// SPDX-License-Identifier: GPL-2.0-or-later 2/* 3 * Asynchronous Compression operations 4 * 5 * Copyright (c) 2016, Intel Corporation 6 * Authors: Weigang Li <weigang.li@intel.com> 7 * Giovanni Cabiddu <giovanni.cabiddu@intel.com> 8 */ 9 10#include <crypto/internal/acompress.h> 11#include <linux/cryptouser.h> 12#include <linux/errno.h> 13#include <linux/kernel.h> 14#include <linux/module.h> 15#include <linux/seq_file.h> 16#include <linux/slab.h> 17#include <linux/string.h> 18#include <net/netlink.h> 19 20#include "compress.h" 21 22struct crypto_scomp; 23 24static const struct crypto_type crypto_acomp_type; 25 26static inline struct acomp_alg *__crypto_acomp_alg(struct crypto_alg *alg) 27{ 28 return container_of(alg, struct acomp_alg, calg.base); 29} 30 31static inline struct acomp_alg *crypto_acomp_alg(struct crypto_acomp *tfm) 32{ 33 return __crypto_acomp_alg(crypto_acomp_tfm(tfm)->__crt_alg); 34} 35 36static int __maybe_unused crypto_acomp_report( 37 struct sk_buff *skb, struct crypto_alg *alg) 38{ 39 struct crypto_report_acomp racomp; 40 41 memset(&racomp, 0, sizeof(racomp)); 42 43 strscpy(racomp.type, "acomp", sizeof(racomp.type)); 44 45 return nla_put(skb, CRYPTOCFGA_REPORT_ACOMP, sizeof(racomp), &racomp); 46} 47 48static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg) 49 __maybe_unused; 50 51static void crypto_acomp_show(struct seq_file *m, struct crypto_alg *alg) 52{ 53 seq_puts(m, "type : acomp\n"); 54} 55 56static void crypto_acomp_exit_tfm(struct crypto_tfm *tfm) 57{ 58 struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm); 59 struct acomp_alg *alg = crypto_acomp_alg(acomp); 60 61 alg->exit(acomp); 62} 63 64static int crypto_acomp_init_tfm(struct crypto_tfm *tfm) 65{ 66 struct crypto_acomp *acomp = __crypto_acomp_tfm(tfm); 67 struct acomp_alg *alg = crypto_acomp_alg(acomp); 68 69 if (tfm->__crt_alg->cra_type != &crypto_acomp_type) 70 return crypto_init_scomp_ops_async(tfm); 71 72 acomp->compress = alg->compress; 73 acomp->decompress = alg->decompress; 74 acomp->dst_free = alg->dst_free; 75 acomp->reqsize = alg->reqsize; 76 77 if (alg->exit) 78 acomp->base.exit = crypto_acomp_exit_tfm; 79 80 if (alg->init) 81 return alg->init(acomp); 82 83 return 0; 84} 85 86static unsigned int crypto_acomp_extsize(struct crypto_alg *alg) 87{ 88 int extsize = crypto_alg_extsize(alg); 89 90 if (alg->cra_type != &crypto_acomp_type) 91 extsize += sizeof(struct crypto_scomp *); 92 93 return extsize; 94} 95 96static const struct crypto_type crypto_acomp_type = { 97 .extsize = crypto_acomp_extsize, 98 .init_tfm = crypto_acomp_init_tfm, 99#ifdef CONFIG_PROC_FS 100 .show = crypto_acomp_show, 101#endif 102#if IS_ENABLED(CONFIG_CRYPTO_USER) 103 .report = crypto_acomp_report, 104#endif 105 .maskclear = ~CRYPTO_ALG_TYPE_MASK, 106 .maskset = CRYPTO_ALG_TYPE_ACOMPRESS_MASK, 107 .type = CRYPTO_ALG_TYPE_ACOMPRESS, 108 .tfmsize = offsetof(struct crypto_acomp, base), 109}; 110 111struct crypto_acomp *crypto_alloc_acomp(const char *alg_name, u32 type, 112 u32 mask) 113{ 114 return crypto_alloc_tfm(alg_name, &crypto_acomp_type, type, mask); 115} 116EXPORT_SYMBOL_GPL(crypto_alloc_acomp); 117 118struct crypto_acomp *crypto_alloc_acomp_node(const char *alg_name, u32 type, 119 u32 mask, int node) 120{ 121 return crypto_alloc_tfm_node(alg_name, &crypto_acomp_type, type, mask, 122 node); 123} 124EXPORT_SYMBOL_GPL(crypto_alloc_acomp_node); 125 126struct acomp_req *acomp_request_alloc(struct crypto_acomp *acomp) 127{ 128 struct crypto_tfm *tfm = crypto_acomp_tfm(acomp); 129 struct acomp_req *req; 130 131 req = __acomp_request_alloc(acomp); 132 if (req && (tfm->__crt_alg->cra_type != &crypto_acomp_type)) 133 return crypto_acomp_scomp_alloc_ctx(req); 134 135 return req; 136} 137EXPORT_SYMBOL_GPL(acomp_request_alloc); 138 139void acomp_request_free(struct acomp_req *req) 140{ 141 struct crypto_acomp *acomp = crypto_acomp_reqtfm(req); 142 struct crypto_tfm *tfm = crypto_acomp_tfm(acomp); 143 144 if (tfm->__crt_alg->cra_type != &crypto_acomp_type) 145 crypto_acomp_scomp_free_ctx(req); 146 147 if (req->flags & CRYPTO_ACOMP_ALLOC_OUTPUT) { 148 acomp->dst_free(req->dst); 149 req->dst = NULL; 150 } 151 152 __acomp_request_free(req); 153} 154EXPORT_SYMBOL_GPL(acomp_request_free); 155 156void comp_prepare_alg(struct comp_alg_common *alg) 157{ 158 struct crypto_alg *base = &alg->base; 159 160 base->cra_flags &= ~CRYPTO_ALG_TYPE_MASK; 161} 162 163int crypto_register_acomp(struct acomp_alg *alg) 164{ 165 struct crypto_alg *base = &alg->calg.base; 166 167 comp_prepare_alg(&alg->calg); 168 169 base->cra_type = &crypto_acomp_type; 170 base->cra_flags |= CRYPTO_ALG_TYPE_ACOMPRESS; 171 172 return crypto_register_alg(base); 173} 174EXPORT_SYMBOL_GPL(crypto_register_acomp); 175 176void crypto_unregister_acomp(struct acomp_alg *alg) 177{ 178 crypto_unregister_alg(&alg->base); 179} 180EXPORT_SYMBOL_GPL(crypto_unregister_acomp); 181 182int crypto_register_acomps(struct acomp_alg *algs, int count) 183{ 184 int i, ret; 185 186 for (i = 0; i < count; i++) { 187 ret = crypto_register_acomp(&algs[i]); 188 if (ret) 189 goto err; 190 } 191 192 return 0; 193 194err: 195 for (--i; i >= 0; --i) 196 crypto_unregister_acomp(&algs[i]); 197 198 return ret; 199} 200EXPORT_SYMBOL_GPL(crypto_register_acomps); 201 202void crypto_unregister_acomps(struct acomp_alg *algs, int count) 203{ 204 int i; 205 206 for (i = count - 1; i >= 0; --i) 207 crypto_unregister_acomp(&algs[i]); 208} 209EXPORT_SYMBOL_GPL(crypto_unregister_acomps); 210 211MODULE_LICENSE("GPL"); 212MODULE_DESCRIPTION("Asynchronous compression type"); 213