mirror of
				git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
				synced 2025-10-31 08:44:41 +00:00 
			
		
		
		
	 c65058b758
			
		
	
	
		c65058b758
		
	
	
	
	
		
			
			Now that all "blkcipher" algorithms have been converted to "skcipher",
remove the blkcipher algorithm type.
The skcipher (symmetric key cipher) algorithm type was introduced a few
years ago to replace both blkcipher and ablkcipher (synchronous and
asynchronous block cipher).  The advantages of skcipher include:
  - A much less confusing name, since none of these algorithm types have
    ever actually been for raw block ciphers, but rather for all
    length-preserving encryption modes including block cipher modes of
    operation, stream ciphers, and other length-preserving modes.
  - It unified blkcipher and ablkcipher into a single algorithm type
    which supports both synchronous and asynchronous implementations.
    Note, blkcipher already operated only on scatterlists, so the fact
    that skcipher does too isn't a regression in functionality.
  - Better type safety by using struct skcipher_alg, struct
    crypto_skcipher, etc. instead of crypto_alg, crypto_tfm, etc.
  - It sometimes simplifies the implementations of algorithms.
Also, the blkcipher API was no longer being tested.
Signed-off-by: Eric Biggers <ebiggers@google.com>
Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
		
	
			
		
			
				
	
	
		
			335 lines
		
	
	
	
		
			9.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			335 lines
		
	
	
	
		
			9.6 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0
 | |
| /*
 | |
|  * Crypto user configuration API.
 | |
|  *
 | |
|  * Copyright (C) 2017-2018 Corentin Labbe <clabbe@baylibre.com>
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #include <linux/crypto.h>
 | |
| #include <linux/cryptouser.h>
 | |
| #include <linux/sched.h>
 | |
| #include <net/netlink.h>
 | |
| #include <net/sock.h>
 | |
| #include <crypto/internal/skcipher.h>
 | |
| #include <crypto/internal/rng.h>
 | |
| #include <crypto/akcipher.h>
 | |
| #include <crypto/kpp.h>
 | |
| #include <crypto/internal/cryptouser.h>
 | |
| 
 | |
| #include "internal.h"
 | |
| 
 | |
| #define null_terminated(x)	(strnlen(x, sizeof(x)) < sizeof(x))
 | |
| 
 | |
| struct crypto_dump_info {
 | |
| 	struct sk_buff *in_skb;
 | |
| 	struct sk_buff *out_skb;
 | |
| 	u32 nlmsg_seq;
 | |
| 	u16 nlmsg_flags;
 | |
| };
 | |
| 
 | |
| static int crypto_report_aead(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_aead raead;
 | |
| 
 | |
| 	memset(&raead, 0, sizeof(raead));
 | |
| 
 | |
| 	strscpy(raead.type, "aead", sizeof(raead.type));
 | |
| 
 | |
| 	raead.stat_encrypt_cnt = atomic64_read(&alg->stats.aead.encrypt_cnt);
 | |
| 	raead.stat_encrypt_tlen = atomic64_read(&alg->stats.aead.encrypt_tlen);
 | |
| 	raead.stat_decrypt_cnt = atomic64_read(&alg->stats.aead.decrypt_cnt);
 | |
| 	raead.stat_decrypt_tlen = atomic64_read(&alg->stats.aead.decrypt_tlen);
 | |
| 	raead.stat_err_cnt = atomic64_read(&alg->stats.aead.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_AEAD, sizeof(raead), &raead);
 | |
| }
 | |
| 
 | |
| static int crypto_report_cipher(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_cipher rcipher;
 | |
| 
 | |
| 	memset(&rcipher, 0, sizeof(rcipher));
 | |
| 
 | |
| 	strscpy(rcipher.type, "cipher", sizeof(rcipher.type));
 | |
| 
 | |
| 	rcipher.stat_encrypt_cnt = atomic64_read(&alg->stats.cipher.encrypt_cnt);
 | |
| 	rcipher.stat_encrypt_tlen = atomic64_read(&alg->stats.cipher.encrypt_tlen);
 | |
| 	rcipher.stat_decrypt_cnt =  atomic64_read(&alg->stats.cipher.decrypt_cnt);
 | |
| 	rcipher.stat_decrypt_tlen = atomic64_read(&alg->stats.cipher.decrypt_tlen);
 | |
| 	rcipher.stat_err_cnt =  atomic64_read(&alg->stats.cipher.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_CIPHER, sizeof(rcipher), &rcipher);
 | |
| }
 | |
| 
 | |
| static int crypto_report_comp(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_compress rcomp;
 | |
| 
 | |
| 	memset(&rcomp, 0, sizeof(rcomp));
 | |
| 
 | |
| 	strscpy(rcomp.type, "compression", sizeof(rcomp.type));
 | |
| 	rcomp.stat_compress_cnt = atomic64_read(&alg->stats.compress.compress_cnt);
 | |
| 	rcomp.stat_compress_tlen = atomic64_read(&alg->stats.compress.compress_tlen);
 | |
| 	rcomp.stat_decompress_cnt = atomic64_read(&alg->stats.compress.decompress_cnt);
 | |
| 	rcomp.stat_decompress_tlen = atomic64_read(&alg->stats.compress.decompress_tlen);
 | |
| 	rcomp.stat_err_cnt = atomic64_read(&alg->stats.compress.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_COMPRESS, sizeof(rcomp), &rcomp);
 | |
| }
 | |
| 
 | |
| static int crypto_report_acomp(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_compress racomp;
 | |
| 
 | |
| 	memset(&racomp, 0, sizeof(racomp));
 | |
| 
 | |
| 	strscpy(racomp.type, "acomp", sizeof(racomp.type));
 | |
| 	racomp.stat_compress_cnt = atomic64_read(&alg->stats.compress.compress_cnt);
 | |
| 	racomp.stat_compress_tlen = atomic64_read(&alg->stats.compress.compress_tlen);
 | |
| 	racomp.stat_decompress_cnt =  atomic64_read(&alg->stats.compress.decompress_cnt);
 | |
| 	racomp.stat_decompress_tlen = atomic64_read(&alg->stats.compress.decompress_tlen);
 | |
| 	racomp.stat_err_cnt = atomic64_read(&alg->stats.compress.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_ACOMP, sizeof(racomp), &racomp);
 | |
| }
 | |
| 
 | |
| static int crypto_report_akcipher(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_akcipher rakcipher;
 | |
| 
 | |
| 	memset(&rakcipher, 0, sizeof(rakcipher));
 | |
| 
 | |
| 	strscpy(rakcipher.type, "akcipher", sizeof(rakcipher.type));
 | |
| 	rakcipher.stat_encrypt_cnt = atomic64_read(&alg->stats.akcipher.encrypt_cnt);
 | |
| 	rakcipher.stat_encrypt_tlen = atomic64_read(&alg->stats.akcipher.encrypt_tlen);
 | |
| 	rakcipher.stat_decrypt_cnt = atomic64_read(&alg->stats.akcipher.decrypt_cnt);
 | |
| 	rakcipher.stat_decrypt_tlen = atomic64_read(&alg->stats.akcipher.decrypt_tlen);
 | |
| 	rakcipher.stat_sign_cnt = atomic64_read(&alg->stats.akcipher.sign_cnt);
 | |
| 	rakcipher.stat_verify_cnt = atomic64_read(&alg->stats.akcipher.verify_cnt);
 | |
| 	rakcipher.stat_err_cnt = atomic64_read(&alg->stats.akcipher.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_AKCIPHER,
 | |
| 		       sizeof(rakcipher), &rakcipher);
 | |
| }
 | |
| 
 | |
| static int crypto_report_kpp(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_kpp rkpp;
 | |
| 
 | |
| 	memset(&rkpp, 0, sizeof(rkpp));
 | |
| 
 | |
| 	strscpy(rkpp.type, "kpp", sizeof(rkpp.type));
 | |
| 
 | |
| 	rkpp.stat_setsecret_cnt = atomic64_read(&alg->stats.kpp.setsecret_cnt);
 | |
| 	rkpp.stat_generate_public_key_cnt = atomic64_read(&alg->stats.kpp.generate_public_key_cnt);
 | |
| 	rkpp.stat_compute_shared_secret_cnt = atomic64_read(&alg->stats.kpp.compute_shared_secret_cnt);
 | |
| 	rkpp.stat_err_cnt = atomic64_read(&alg->stats.kpp.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_KPP, sizeof(rkpp), &rkpp);
 | |
| }
 | |
| 
 | |
| static int crypto_report_ahash(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_hash rhash;
 | |
| 
 | |
| 	memset(&rhash, 0, sizeof(rhash));
 | |
| 
 | |
| 	strscpy(rhash.type, "ahash", sizeof(rhash.type));
 | |
| 
 | |
| 	rhash.stat_hash_cnt = atomic64_read(&alg->stats.hash.hash_cnt);
 | |
| 	rhash.stat_hash_tlen = atomic64_read(&alg->stats.hash.hash_tlen);
 | |
| 	rhash.stat_err_cnt = atomic64_read(&alg->stats.hash.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(rhash), &rhash);
 | |
| }
 | |
| 
 | |
| static int crypto_report_shash(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_hash rhash;
 | |
| 
 | |
| 	memset(&rhash, 0, sizeof(rhash));
 | |
| 
 | |
| 	strscpy(rhash.type, "shash", sizeof(rhash.type));
 | |
| 
 | |
| 	rhash.stat_hash_cnt =  atomic64_read(&alg->stats.hash.hash_cnt);
 | |
| 	rhash.stat_hash_tlen = atomic64_read(&alg->stats.hash.hash_tlen);
 | |
| 	rhash.stat_err_cnt = atomic64_read(&alg->stats.hash.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_HASH, sizeof(rhash), &rhash);
 | |
| }
 | |
| 
 | |
| static int crypto_report_rng(struct sk_buff *skb, struct crypto_alg *alg)
 | |
| {
 | |
| 	struct crypto_stat_rng rrng;
 | |
| 
 | |
| 	memset(&rrng, 0, sizeof(rrng));
 | |
| 
 | |
| 	strscpy(rrng.type, "rng", sizeof(rrng.type));
 | |
| 
 | |
| 	rrng.stat_generate_cnt = atomic64_read(&alg->stats.rng.generate_cnt);
 | |
| 	rrng.stat_generate_tlen = atomic64_read(&alg->stats.rng.generate_tlen);
 | |
| 	rrng.stat_seed_cnt = atomic64_read(&alg->stats.rng.seed_cnt);
 | |
| 	rrng.stat_err_cnt = atomic64_read(&alg->stats.rng.err_cnt);
 | |
| 
 | |
| 	return nla_put(skb, CRYPTOCFGA_STAT_RNG, sizeof(rrng), &rrng);
 | |
| }
 | |
| 
 | |
| static int crypto_reportstat_one(struct crypto_alg *alg,
 | |
| 				 struct crypto_user_alg *ualg,
 | |
| 				 struct sk_buff *skb)
 | |
| {
 | |
| 	memset(ualg, 0, sizeof(*ualg));
 | |
| 
 | |
| 	strscpy(ualg->cru_name, alg->cra_name, sizeof(ualg->cru_name));
 | |
| 	strscpy(ualg->cru_driver_name, alg->cra_driver_name,
 | |
| 		sizeof(ualg->cru_driver_name));
 | |
| 	strscpy(ualg->cru_module_name, module_name(alg->cra_module),
 | |
| 		sizeof(ualg->cru_module_name));
 | |
| 
 | |
| 	ualg->cru_type = 0;
 | |
| 	ualg->cru_mask = 0;
 | |
| 	ualg->cru_flags = alg->cra_flags;
 | |
| 	ualg->cru_refcnt = refcount_read(&alg->cra_refcnt);
 | |
| 
 | |
| 	if (nla_put_u32(skb, CRYPTOCFGA_PRIORITY_VAL, alg->cra_priority))
 | |
| 		goto nla_put_failure;
 | |
| 	if (alg->cra_flags & CRYPTO_ALG_LARVAL) {
 | |
| 		struct crypto_stat_larval rl;
 | |
| 
 | |
| 		memset(&rl, 0, sizeof(rl));
 | |
| 		strscpy(rl.type, "larval", sizeof(rl.type));
 | |
| 		if (nla_put(skb, CRYPTOCFGA_STAT_LARVAL, sizeof(rl), &rl))
 | |
| 			goto nla_put_failure;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	switch (alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL)) {
 | |
| 	case CRYPTO_ALG_TYPE_AEAD:
 | |
| 		if (crypto_report_aead(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_SKCIPHER:
 | |
| 		if (crypto_report_cipher(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_CIPHER:
 | |
| 		if (crypto_report_cipher(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_COMPRESS:
 | |
| 		if (crypto_report_comp(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_ACOMPRESS:
 | |
| 		if (crypto_report_acomp(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_SCOMPRESS:
 | |
| 		if (crypto_report_acomp(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_AKCIPHER:
 | |
| 		if (crypto_report_akcipher(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_KPP:
 | |
| 		if (crypto_report_kpp(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_AHASH:
 | |
| 		if (crypto_report_ahash(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_HASH:
 | |
| 		if (crypto_report_shash(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	case CRYPTO_ALG_TYPE_RNG:
 | |
| 		if (crypto_report_rng(skb, alg))
 | |
| 			goto nla_put_failure;
 | |
| 		break;
 | |
| 	default:
 | |
| 		pr_err("ERROR: Unhandled alg %d in %s\n",
 | |
| 		       alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL),
 | |
| 		       __func__);
 | |
| 	}
 | |
| 
 | |
| out:
 | |
| 	return 0;
 | |
| 
 | |
| nla_put_failure:
 | |
| 	return -EMSGSIZE;
 | |
| }
 | |
| 
 | |
| static int crypto_reportstat_alg(struct crypto_alg *alg,
 | |
| 				 struct crypto_dump_info *info)
 | |
| {
 | |
| 	struct sk_buff *in_skb = info->in_skb;
 | |
| 	struct sk_buff *skb = info->out_skb;
 | |
| 	struct nlmsghdr *nlh;
 | |
| 	struct crypto_user_alg *ualg;
 | |
| 	int err = 0;
 | |
| 
 | |
| 	nlh = nlmsg_put(skb, NETLINK_CB(in_skb).portid, info->nlmsg_seq,
 | |
| 			CRYPTO_MSG_GETSTAT, sizeof(*ualg), info->nlmsg_flags);
 | |
| 	if (!nlh) {
 | |
| 		err = -EMSGSIZE;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	ualg = nlmsg_data(nlh);
 | |
| 
 | |
| 	err = crypto_reportstat_one(alg, ualg, skb);
 | |
| 	if (err) {
 | |
| 		nlmsg_cancel(skb, nlh);
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	nlmsg_end(skb, nlh);
 | |
| 
 | |
| out:
 | |
| 	return err;
 | |
| }
 | |
| 
 | |
| int crypto_reportstat(struct sk_buff *in_skb, struct nlmsghdr *in_nlh,
 | |
| 		      struct nlattr **attrs)
 | |
| {
 | |
| 	struct net *net = sock_net(in_skb->sk);
 | |
| 	struct crypto_user_alg *p = nlmsg_data(in_nlh);
 | |
| 	struct crypto_alg *alg;
 | |
| 	struct sk_buff *skb;
 | |
| 	struct crypto_dump_info info;
 | |
| 	int err;
 | |
| 
 | |
| 	if (!null_terminated(p->cru_name) || !null_terminated(p->cru_driver_name))
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	alg = crypto_alg_match(p, 0);
 | |
| 	if (!alg)
 | |
| 		return -ENOENT;
 | |
| 
 | |
| 	err = -ENOMEM;
 | |
| 	skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_ATOMIC);
 | |
| 	if (!skb)
 | |
| 		goto drop_alg;
 | |
| 
 | |
| 	info.in_skb = in_skb;
 | |
| 	info.out_skb = skb;
 | |
| 	info.nlmsg_seq = in_nlh->nlmsg_seq;
 | |
| 	info.nlmsg_flags = 0;
 | |
| 
 | |
| 	err = crypto_reportstat_alg(alg, &info);
 | |
| 
 | |
| drop_alg:
 | |
| 	crypto_mod_put(alg);
 | |
| 
 | |
| 	if (err) {
 | |
| 		kfree_skb(skb);
 | |
| 		return err;
 | |
| 	}
 | |
| 
 | |
| 	return nlmsg_unicast(net->crypto_nlsk, skb, NETLINK_CB(in_skb).portid);
 | |
| }
 | |
| 
 | |
| MODULE_LICENSE("GPL");
 |