| // SPDX-License-Identifier: GPL-2.0 | 
 | /* | 
 |  * Crypto user configuration API. | 
 |  * | 
 |  * Copyright (C) 2017-2018 Corentin Labbe <clabbe@baylibre.com> | 
 |  * | 
 |  */ | 
 |  | 
 | #include <linux/crypto.h> | 
 | #include <linux/cryptouser.h> | 
 | #include <linux/sched.h> | 
 | #include <net/netlink.h> | 
 | #include <crypto/internal/skcipher.h> | 
 | #include <crypto/internal/rng.h> | 
 | #include <crypto/akcipher.h> | 
 | #include <crypto/kpp.h> | 
 | #include <crypto/internal/cryptouser.h> | 
 |  | 
 | #include "internal.h" | 
 |  | 
 | #define null_terminated(x)	(strnlen(x, sizeof(x)) < sizeof(x)) | 
 |  | 
 | static DEFINE_MUTEX(crypto_cfg_mutex); | 
 |  | 
 | extern struct sock *crypto_nlsk; | 
 |  | 
 | struct crypto_dump_info { | 
 | 	struct sk_buff *in_skb; | 
 | 	struct sk_buff *out_skb; | 
 | 	u32 nlmsg_seq; | 
 | 	u16 nlmsg_flags; | 
 | }; | 
 |  | 
 | static int crypto_report_aead(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat raead; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&raead, 0, sizeof(raead)); | 
 |  | 
 | 	strncpy(raead.type, "aead", sizeof(raead.type)); | 
 |  | 
 | 	v32 = atomic_read(&alg->encrypt_cnt); | 
 | 	raead.stat_encrypt_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->encrypt_tlen); | 
 | 	raead.stat_encrypt_tlen = v64; | 
 | 	v32 = atomic_read(&alg->decrypt_cnt); | 
 | 	raead.stat_decrypt_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->decrypt_tlen); | 
 | 	raead.stat_decrypt_tlen = v64; | 
 | 	v32 = atomic_read(&alg->aead_err_cnt); | 
 | 	raead.stat_aead_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_AEAD, | 
 | 		    sizeof(struct crypto_stat), &raead)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_cipher(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat rcipher; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&rcipher, 0, sizeof(rcipher)); | 
 |  | 
 | 	strlcpy(rcipher.type, "cipher", sizeof(rcipher.type)); | 
 |  | 
 | 	v32 = atomic_read(&alg->encrypt_cnt); | 
 | 	rcipher.stat_encrypt_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->encrypt_tlen); | 
 | 	rcipher.stat_encrypt_tlen = v64; | 
 | 	v32 = atomic_read(&alg->decrypt_cnt); | 
 | 	rcipher.stat_decrypt_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->decrypt_tlen); | 
 | 	rcipher.stat_decrypt_tlen = v64; | 
 | 	v32 = atomic_read(&alg->cipher_err_cnt); | 
 | 	rcipher.stat_cipher_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_CIPHER, | 
 | 		    sizeof(struct crypto_stat), &rcipher)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_comp(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat rcomp; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&rcomp, 0, sizeof(rcomp)); | 
 |  | 
 | 	strlcpy(rcomp.type, "compression", sizeof(rcomp.type)); | 
 | 	v32 = atomic_read(&alg->compress_cnt); | 
 | 	rcomp.stat_compress_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->compress_tlen); | 
 | 	rcomp.stat_compress_tlen = v64; | 
 | 	v32 = atomic_read(&alg->decompress_cnt); | 
 | 	rcomp.stat_decompress_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->decompress_tlen); | 
 | 	rcomp.stat_decompress_tlen = v64; | 
 | 	v32 = atomic_read(&alg->cipher_err_cnt); | 
 | 	rcomp.stat_compress_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_COMPRESS, | 
 | 		    sizeof(struct crypto_stat), &rcomp)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_acomp(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat racomp; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&racomp, 0, sizeof(racomp)); | 
 |  | 
 | 	strlcpy(racomp.type, "acomp", sizeof(racomp.type)); | 
 | 	v32 = atomic_read(&alg->compress_cnt); | 
 | 	racomp.stat_compress_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->compress_tlen); | 
 | 	racomp.stat_compress_tlen = v64; | 
 | 	v32 = atomic_read(&alg->decompress_cnt); | 
 | 	racomp.stat_decompress_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->decompress_tlen); | 
 | 	racomp.stat_decompress_tlen = v64; | 
 | 	v32 = atomic_read(&alg->cipher_err_cnt); | 
 | 	racomp.stat_compress_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_ACOMP, | 
 | 		    sizeof(struct crypto_stat), &racomp)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_akcipher(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat rakcipher; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&rakcipher, 0, sizeof(rakcipher)); | 
 |  | 
 | 	strncpy(rakcipher.type, "akcipher", sizeof(rakcipher.type)); | 
 | 	v32 = atomic_read(&alg->encrypt_cnt); | 
 | 	rakcipher.stat_encrypt_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->encrypt_tlen); | 
 | 	rakcipher.stat_encrypt_tlen = v64; | 
 | 	v32 = atomic_read(&alg->decrypt_cnt); | 
 | 	rakcipher.stat_decrypt_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->decrypt_tlen); | 
 | 	rakcipher.stat_decrypt_tlen = v64; | 
 | 	v32 = atomic_read(&alg->sign_cnt); | 
 | 	rakcipher.stat_sign_cnt = v32; | 
 | 	v32 = atomic_read(&alg->verify_cnt); | 
 | 	rakcipher.stat_verify_cnt = v32; | 
 | 	v32 = atomic_read(&alg->akcipher_err_cnt); | 
 | 	rakcipher.stat_akcipher_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_AKCIPHER, | 
 | 		    sizeof(struct crypto_stat), &rakcipher)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_kpp(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat rkpp; | 
 | 	u32 v; | 
 |  | 
 | 	memset(&rkpp, 0, sizeof(rkpp)); | 
 |  | 
 | 	strlcpy(rkpp.type, "kpp", sizeof(rkpp.type)); | 
 |  | 
 | 	v = atomic_read(&alg->setsecret_cnt); | 
 | 	rkpp.stat_setsecret_cnt = v; | 
 | 	v = atomic_read(&alg->generate_public_key_cnt); | 
 | 	rkpp.stat_generate_public_key_cnt = v; | 
 | 	v = atomic_read(&alg->compute_shared_secret_cnt); | 
 | 	rkpp.stat_compute_shared_secret_cnt = v; | 
 | 	v = atomic_read(&alg->kpp_err_cnt); | 
 | 	rkpp.stat_kpp_err_cnt = v; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_KPP, | 
 | 		    sizeof(struct crypto_stat), &rkpp)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_ahash(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat rhash; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&rhash, 0, sizeof(rhash)); | 
 |  | 
 | 	strncpy(rhash.type, "ahash", sizeof(rhash.type)); | 
 |  | 
 | 	v32 = atomic_read(&alg->hash_cnt); | 
 | 	rhash.stat_hash_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->hash_tlen); | 
 | 	rhash.stat_hash_tlen = v64; | 
 | 	v32 = atomic_read(&alg->hash_err_cnt); | 
 | 	rhash.stat_hash_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_HASH, | 
 | 		    sizeof(struct crypto_stat), &rhash)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_shash(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat rhash; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&rhash, 0, sizeof(rhash)); | 
 |  | 
 | 	strncpy(rhash.type, "shash", sizeof(rhash.type)); | 
 |  | 
 | 	v32 = atomic_read(&alg->hash_cnt); | 
 | 	rhash.stat_hash_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->hash_tlen); | 
 | 	rhash.stat_hash_tlen = v64; | 
 | 	v32 = atomic_read(&alg->hash_err_cnt); | 
 | 	rhash.stat_hash_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_HASH, | 
 | 		    sizeof(struct crypto_stat), &rhash)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_report_rng(struct sk_buff *skb, struct crypto_alg *alg) | 
 | { | 
 | 	struct crypto_stat rrng; | 
 | 	u64 v64; | 
 | 	u32 v32; | 
 |  | 
 | 	memset(&rrng, 0, sizeof(rrng)); | 
 |  | 
 | 	strncpy(rrng.type, "rng", sizeof(rrng.type)); | 
 |  | 
 | 	v32 = atomic_read(&alg->generate_cnt); | 
 | 	rrng.stat_generate_cnt = v32; | 
 | 	v64 = atomic64_read(&alg->generate_tlen); | 
 | 	rrng.stat_generate_tlen = v64; | 
 | 	v32 = atomic_read(&alg->seed_cnt); | 
 | 	rrng.stat_seed_cnt = v32; | 
 | 	v32 = atomic_read(&alg->hash_err_cnt); | 
 | 	rrng.stat_rng_err_cnt = v32; | 
 |  | 
 | 	if (nla_put(skb, CRYPTOCFGA_STAT_RNG, | 
 | 		    sizeof(struct crypto_stat), &rrng)) | 
 | 		goto nla_put_failure; | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_reportstat_one(struct crypto_alg *alg, | 
 | 				 struct crypto_user_alg *ualg, | 
 | 				 struct sk_buff *skb) | 
 | { | 
 | 	memset(ualg, 0, sizeof(*ualg)); | 
 |  | 
 | 	strlcpy(ualg->cru_name, alg->cra_name, sizeof(ualg->cru_name)); | 
 | 	strlcpy(ualg->cru_driver_name, alg->cra_driver_name, | 
 | 		sizeof(ualg->cru_driver_name)); | 
 | 	strlcpy(ualg->cru_module_name, module_name(alg->cra_module), | 
 | 		sizeof(ualg->cru_module_name)); | 
 |  | 
 | 	ualg->cru_type = 0; | 
 | 	ualg->cru_mask = 0; | 
 | 	ualg->cru_flags = alg->cra_flags; | 
 | 	ualg->cru_refcnt = refcount_read(&alg->cra_refcnt); | 
 |  | 
 | 	if (nla_put_u32(skb, CRYPTOCFGA_PRIORITY_VAL, alg->cra_priority)) | 
 | 		goto nla_put_failure; | 
 | 	if (alg->cra_flags & CRYPTO_ALG_LARVAL) { | 
 | 		struct crypto_stat rl; | 
 |  | 
 | 		memset(&rl, 0, sizeof(rl)); | 
 | 		strlcpy(rl.type, "larval", sizeof(rl.type)); | 
 | 		if (nla_put(skb, CRYPTOCFGA_STAT_LARVAL, | 
 | 			    sizeof(struct crypto_stat), &rl)) | 
 | 			goto nla_put_failure; | 
 | 		goto out; | 
 | 	} | 
 |  | 
 | 	switch (alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL)) { | 
 | 	case CRYPTO_ALG_TYPE_AEAD: | 
 | 		if (crypto_report_aead(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_SKCIPHER: | 
 | 		if (crypto_report_cipher(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_BLKCIPHER: | 
 | 		if (crypto_report_cipher(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_CIPHER: | 
 | 		if (crypto_report_cipher(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_COMPRESS: | 
 | 		if (crypto_report_comp(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_ACOMPRESS: | 
 | 		if (crypto_report_acomp(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_SCOMPRESS: | 
 | 		if (crypto_report_acomp(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_AKCIPHER: | 
 | 		if (crypto_report_akcipher(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_KPP: | 
 | 		if (crypto_report_kpp(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_AHASH: | 
 | 		if (crypto_report_ahash(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_HASH: | 
 | 		if (crypto_report_shash(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	case CRYPTO_ALG_TYPE_RNG: | 
 | 		if (crypto_report_rng(skb, alg)) | 
 | 			goto nla_put_failure; | 
 | 		break; | 
 | 	default: | 
 | 		pr_err("ERROR: Unhandled alg %d in %s\n", | 
 | 		       alg->cra_flags & (CRYPTO_ALG_TYPE_MASK | CRYPTO_ALG_LARVAL), | 
 | 		       __func__); | 
 | 	} | 
 |  | 
 | out: | 
 | 	return 0; | 
 |  | 
 | nla_put_failure: | 
 | 	return -EMSGSIZE; | 
 | } | 
 |  | 
 | static int crypto_reportstat_alg(struct crypto_alg *alg, | 
 | 				 struct crypto_dump_info *info) | 
 | { | 
 | 	struct sk_buff *in_skb = info->in_skb; | 
 | 	struct sk_buff *skb = info->out_skb; | 
 | 	struct nlmsghdr *nlh; | 
 | 	struct crypto_user_alg *ualg; | 
 | 	int err = 0; | 
 |  | 
 | 	nlh = nlmsg_put(skb, NETLINK_CB(in_skb).portid, info->nlmsg_seq, | 
 | 			CRYPTO_MSG_GETSTAT, sizeof(*ualg), info->nlmsg_flags); | 
 | 	if (!nlh) { | 
 | 		err = -EMSGSIZE; | 
 | 		goto out; | 
 | 	} | 
 |  | 
 | 	ualg = nlmsg_data(nlh); | 
 |  | 
 | 	err = crypto_reportstat_one(alg, ualg, skb); | 
 | 	if (err) { | 
 | 		nlmsg_cancel(skb, nlh); | 
 | 		goto out; | 
 | 	} | 
 |  | 
 | 	nlmsg_end(skb, nlh); | 
 |  | 
 | out: | 
 | 	return err; | 
 | } | 
 |  | 
 | int crypto_reportstat(struct sk_buff *in_skb, struct nlmsghdr *in_nlh, | 
 | 		      struct nlattr **attrs) | 
 | { | 
 | 	struct crypto_user_alg *p = nlmsg_data(in_nlh); | 
 | 	struct crypto_alg *alg; | 
 | 	struct sk_buff *skb; | 
 | 	struct crypto_dump_info info; | 
 | 	int err; | 
 |  | 
 | 	if (!null_terminated(p->cru_name) || !null_terminated(p->cru_driver_name)) | 
 | 		return -EINVAL; | 
 |  | 
 | 	alg = crypto_alg_match(p, 0); | 
 | 	if (!alg) | 
 | 		return -ENOENT; | 
 |  | 
 | 	err = -ENOMEM; | 
 | 	skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_ATOMIC); | 
 | 	if (!skb) | 
 | 		goto drop_alg; | 
 |  | 
 | 	info.in_skb = in_skb; | 
 | 	info.out_skb = skb; | 
 | 	info.nlmsg_seq = in_nlh->nlmsg_seq; | 
 | 	info.nlmsg_flags = 0; | 
 |  | 
 | 	err = crypto_reportstat_alg(alg, &info); | 
 |  | 
 | drop_alg: | 
 | 	crypto_mod_put(alg); | 
 |  | 
 | 	if (err) | 
 | 		return err; | 
 |  | 
 | 	return nlmsg_unicast(crypto_nlsk, skb, NETLINK_CB(in_skb).portid); | 
 | } | 
 |  | 
 | int crypto_dump_reportstat(struct sk_buff *skb, struct netlink_callback *cb) | 
 | { | 
 | 	struct crypto_alg *alg; | 
 | 	struct crypto_dump_info info; | 
 | 	int err; | 
 |  | 
 | 	if (cb->args[0]) | 
 | 		goto out; | 
 |  | 
 | 	cb->args[0] = 1; | 
 |  | 
 | 	info.in_skb = cb->skb; | 
 | 	info.out_skb = skb; | 
 | 	info.nlmsg_seq = cb->nlh->nlmsg_seq; | 
 | 	info.nlmsg_flags = NLM_F_MULTI; | 
 |  | 
 | 	list_for_each_entry(alg, &crypto_alg_list, cra_list) { | 
 | 		err = crypto_reportstat_alg(alg, &info); | 
 | 		if (err) | 
 | 			goto out_err; | 
 | 	} | 
 |  | 
 | out: | 
 | 	return skb->len; | 
 | out_err: | 
 | 	return err; | 
 | } | 
 |  | 
 | int crypto_dump_reportstat_done(struct netlink_callback *cb) | 
 | { | 
 | 	return 0; | 
 | } | 
 |  | 
 | MODULE_LICENSE("GPL"); |