mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-09-18 22:14:16 +00:00
crypto: sha256 - support arch-optimized lib and expose through shash
As has been done for various other algorithms, rework the design of the SHA-256 library to support arch-optimized implementations, and make crypto/sha256.c expose both generic and arch-optimized shash algorithms that wrap the library functions. This allows users of the SHA-256 library functions to take advantage of the arch-optimized code, and this makes it much simpler to integrate SHA-256 for each architecture. Note that sha256_base.h is not used in the new design. It will be removed once all the architecture-specific code has been updated. Move the generic block function into its own module to avoid a circular dependency from libsha256.ko => sha256-$ARCH.ko => libsha256.ko. Signed-off-by: Eric Biggers <ebiggers@google.com> Add export and import functions to maintain existing export format. Signed-off-by: Herbert Xu <herbert@gondor.apana.org.au>
This commit is contained in:
parent
10a6d72ea3
commit
950e5c8411
11 changed files with 529 additions and 227 deletions
|
@ -982,6 +982,7 @@ config CRYPTO_SHA256
|
|||
tristate "SHA-224 and SHA-256"
|
||||
select CRYPTO_HASH
|
||||
select CRYPTO_LIB_SHA256
|
||||
select CRYPTO_LIB_SHA256_GENERIC
|
||||
help
|
||||
SHA-224 and SHA-256 secure hash algorithms (FIPS 180, ISO/IEC 10118-3)
|
||||
|
||||
|
|
|
@ -76,7 +76,8 @@ obj-$(CONFIG_CRYPTO_MD4) += md4.o
|
|||
obj-$(CONFIG_CRYPTO_MD5) += md5.o
|
||||
obj-$(CONFIG_CRYPTO_RMD160) += rmd160.o
|
||||
obj-$(CONFIG_CRYPTO_SHA1) += sha1_generic.o
|
||||
obj-$(CONFIG_CRYPTO_SHA256) += sha256_generic.o
|
||||
obj-$(CONFIG_CRYPTO_SHA256) += sha256.o
|
||||
CFLAGS_sha256.o += -DARCH=$(ARCH)
|
||||
obj-$(CONFIG_CRYPTO_SHA512) += sha512_generic.o
|
||||
obj-$(CONFIG_CRYPTO_SHA3) += sha3_generic.o
|
||||
obj-$(CONFIG_CRYPTO_SM3_GENERIC) += sm3_generic.o
|
||||
|
|
243
crypto/sha256.c
Normal file
243
crypto/sha256.c
Normal file
|
@ -0,0 +1,243 @@
|
|||
// SPDX-License-Identifier: GPL-2.0-or-later
|
||||
/*
|
||||
* Crypto API wrapper for the SHA-256 and SHA-224 library functions
|
||||
*
|
||||
* Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com>
|
||||
* Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk>
|
||||
* Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
|
||||
* SHA224 Support Copyright 2007 Intel Corporation <jonathan.lynch@intel.com>
|
||||
*/
|
||||
#include <crypto/internal/hash.h>
|
||||
#include <crypto/internal/sha2.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/module.h>
|
||||
|
||||
const u8 sha224_zero_message_hash[SHA224_DIGEST_SIZE] = {
|
||||
0xd1, 0x4a, 0x02, 0x8c, 0x2a, 0x3a, 0x2b, 0xc9, 0x47,
|
||||
0x61, 0x02, 0xbb, 0x28, 0x82, 0x34, 0xc4, 0x15, 0xa2,
|
||||
0xb0, 0x1f, 0x82, 0x8e, 0xa6, 0x2a, 0xc5, 0xb3, 0xe4,
|
||||
0x2f
|
||||
};
|
||||
EXPORT_SYMBOL_GPL(sha224_zero_message_hash);
|
||||
|
||||
const u8 sha256_zero_message_hash[SHA256_DIGEST_SIZE] = {
|
||||
0xe3, 0xb0, 0xc4, 0x42, 0x98, 0xfc, 0x1c, 0x14,
|
||||
0x9a, 0xfb, 0xf4, 0xc8, 0x99, 0x6f, 0xb9, 0x24,
|
||||
0x27, 0xae, 0x41, 0xe4, 0x64, 0x9b, 0x93, 0x4c,
|
||||
0xa4, 0x95, 0x99, 0x1b, 0x78, 0x52, 0xb8, 0x55
|
||||
};
|
||||
EXPORT_SYMBOL_GPL(sha256_zero_message_hash);
|
||||
|
||||
static int crypto_sha256_init(struct shash_desc *desc)
|
||||
{
|
||||
sha256_init(shash_desc_ctx(desc));
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_update_generic(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len)
|
||||
{
|
||||
sha256_update_generic(shash_desc_ctx(desc), data, len);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_update_arch(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len)
|
||||
{
|
||||
sha256_update(shash_desc_ctx(desc), data, len);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_final_generic(struct shash_desc *desc, u8 *out)
|
||||
{
|
||||
sha256_final_generic(shash_desc_ctx(desc), out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_final_arch(struct shash_desc *desc, u8 *out)
|
||||
{
|
||||
sha256_final(shash_desc_ctx(desc), out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_finup_generic(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len, u8 *out)
|
||||
{
|
||||
struct sha256_state *sctx = shash_desc_ctx(desc);
|
||||
|
||||
sha256_update_generic(sctx, data, len);
|
||||
sha256_final_generic(sctx, out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_finup_arch(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len, u8 *out)
|
||||
{
|
||||
struct sha256_state *sctx = shash_desc_ctx(desc);
|
||||
|
||||
sha256_update(sctx, data, len);
|
||||
sha256_final(sctx, out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_digest_generic(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len, u8 *out)
|
||||
{
|
||||
struct sha256_state *sctx = shash_desc_ctx(desc);
|
||||
|
||||
sha256_init(sctx);
|
||||
sha256_update_generic(sctx, data, len);
|
||||
sha256_final_generic(sctx, out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_digest_arch(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len, u8 *out)
|
||||
{
|
||||
sha256(data, len, out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha224_init(struct shash_desc *desc)
|
||||
{
|
||||
sha224_init(shash_desc_ctx(desc));
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha224_final_generic(struct shash_desc *desc, u8 *out)
|
||||
{
|
||||
sha224_final_generic(shash_desc_ctx(desc), out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha224_final_arch(struct shash_desc *desc, u8 *out)
|
||||
{
|
||||
sha224_final(shash_desc_ctx(desc), out);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_import_lib(struct shash_desc *desc, const void *in)
|
||||
{
|
||||
struct sha256_state *sctx = shash_desc_ctx(desc);
|
||||
const u8 *p = in;
|
||||
|
||||
memcpy(sctx, p, sizeof(*sctx));
|
||||
p += sizeof(*sctx);
|
||||
sctx->count += *p;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int crypto_sha256_export_lib(struct shash_desc *desc, void *out)
|
||||
{
|
||||
struct sha256_state *sctx0 = shash_desc_ctx(desc);
|
||||
struct sha256_state sctx = *sctx0;
|
||||
unsigned int partial;
|
||||
u8 *p = out;
|
||||
|
||||
partial = sctx.count % SHA256_BLOCK_SIZE;
|
||||
sctx.count -= partial;
|
||||
memcpy(p, &sctx, sizeof(sctx));
|
||||
p += sizeof(sctx);
|
||||
*p = partial;
|
||||
return 0;
|
||||
}
|
||||
|
||||
static struct shash_alg algs[] = {
|
||||
{
|
||||
.base.cra_name = "sha256",
|
||||
.base.cra_driver_name = "sha256-generic",
|
||||
.base.cra_priority = 100,
|
||||
.base.cra_blocksize = SHA256_BLOCK_SIZE,
|
||||
.base.cra_module = THIS_MODULE,
|
||||
.digestsize = SHA256_DIGEST_SIZE,
|
||||
.init = crypto_sha256_init,
|
||||
.update = crypto_sha256_update_generic,
|
||||
.final = crypto_sha256_final_generic,
|
||||
.finup = crypto_sha256_finup_generic,
|
||||
.digest = crypto_sha256_digest_generic,
|
||||
.descsize = sizeof(struct sha256_state),
|
||||
.statesize = sizeof(struct crypto_sha256_state) +
|
||||
SHA256_BLOCK_SIZE + 1,
|
||||
.import = crypto_sha256_import_lib,
|
||||
.export = crypto_sha256_export_lib,
|
||||
},
|
||||
{
|
||||
.base.cra_name = "sha224",
|
||||
.base.cra_driver_name = "sha224-generic",
|
||||
.base.cra_priority = 100,
|
||||
.base.cra_blocksize = SHA224_BLOCK_SIZE,
|
||||
.base.cra_module = THIS_MODULE,
|
||||
.digestsize = SHA224_DIGEST_SIZE,
|
||||
.init = crypto_sha224_init,
|
||||
.update = crypto_sha256_update_generic,
|
||||
.final = crypto_sha224_final_generic,
|
||||
.descsize = sizeof(struct sha256_state),
|
||||
.statesize = sizeof(struct crypto_sha256_state) +
|
||||
SHA256_BLOCK_SIZE + 1,
|
||||
.import = crypto_sha256_import_lib,
|
||||
.export = crypto_sha256_export_lib,
|
||||
},
|
||||
{
|
||||
.base.cra_name = "sha256",
|
||||
.base.cra_driver_name = "sha256-" __stringify(ARCH),
|
||||
.base.cra_priority = 300,
|
||||
.base.cra_blocksize = SHA256_BLOCK_SIZE,
|
||||
.base.cra_module = THIS_MODULE,
|
||||
.digestsize = SHA256_DIGEST_SIZE,
|
||||
.init = crypto_sha256_init,
|
||||
.update = crypto_sha256_update_arch,
|
||||
.final = crypto_sha256_final_arch,
|
||||
.finup = crypto_sha256_finup_arch,
|
||||
.digest = crypto_sha256_digest_arch,
|
||||
.descsize = sizeof(struct sha256_state),
|
||||
.statesize = sizeof(struct crypto_sha256_state) +
|
||||
SHA256_BLOCK_SIZE + 1,
|
||||
.import = crypto_sha256_import_lib,
|
||||
.export = crypto_sha256_export_lib,
|
||||
},
|
||||
{
|
||||
.base.cra_name = "sha224",
|
||||
.base.cra_driver_name = "sha224-" __stringify(ARCH),
|
||||
.base.cra_priority = 300,
|
||||
.base.cra_blocksize = SHA224_BLOCK_SIZE,
|
||||
.base.cra_module = THIS_MODULE,
|
||||
.digestsize = SHA224_DIGEST_SIZE,
|
||||
.init = crypto_sha224_init,
|
||||
.update = crypto_sha256_update_arch,
|
||||
.final = crypto_sha224_final_arch,
|
||||
.descsize = sizeof(struct sha256_state),
|
||||
.statesize = sizeof(struct crypto_sha256_state) +
|
||||
SHA256_BLOCK_SIZE + 1,
|
||||
.import = crypto_sha256_import_lib,
|
||||
.export = crypto_sha256_export_lib,
|
||||
},
|
||||
};
|
||||
|
||||
static unsigned int num_algs;
|
||||
|
||||
static int __init crypto_sha256_mod_init(void)
|
||||
{
|
||||
/* register the arch flavours only if they differ from generic */
|
||||
num_algs = ARRAY_SIZE(algs);
|
||||
BUILD_BUG_ON(ARRAY_SIZE(algs) % 2 != 0);
|
||||
if (!sha256_is_arch_optimized())
|
||||
num_algs /= 2;
|
||||
return crypto_register_shashes(algs, ARRAY_SIZE(algs));
|
||||
}
|
||||
subsys_initcall(crypto_sha256_mod_init);
|
||||
|
||||
static void __exit crypto_sha256_mod_exit(void)
|
||||
{
|
||||
crypto_unregister_shashes(algs, num_algs);
|
||||
}
|
||||
module_exit(crypto_sha256_mod_exit);
|
||||
|
||||
MODULE_LICENSE("GPL");
|
||||
MODULE_DESCRIPTION("Crypto API wrapper for the SHA-256 and SHA-224 library functions");
|
||||
|
||||
MODULE_ALIAS_CRYPTO("sha256");
|
||||
MODULE_ALIAS_CRYPTO("sha256-generic");
|
||||
MODULE_ALIAS_CRYPTO("sha256-" __stringify(ARCH));
|
||||
MODULE_ALIAS_CRYPTO("sha224");
|
||||
MODULE_ALIAS_CRYPTO("sha224-generic");
|
||||
MODULE_ALIAS_CRYPTO("sha224-" __stringify(ARCH));
|
|
@ -1,102 +0,0 @@
|
|||
// SPDX-License-Identifier: GPL-2.0-or-later
|
||||
/*
|
||||
* Crypto API wrapper for the generic SHA256 code from lib/crypto/sha256.c
|
||||
*
|
||||
* Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com>
|
||||
* Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk>
|
||||
* Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
|
||||
* SHA224 Support Copyright 2007 Intel Corporation <jonathan.lynch@intel.com>
|
||||
*/
|
||||
#include <crypto/internal/hash.h>
|
||||
#include <crypto/sha2.h>
|
||||
#include <crypto/sha256_base.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/module.h>
|
||||
|
||||
const u8 sha224_zero_message_hash[SHA224_DIGEST_SIZE] = {
|
||||
0xd1, 0x4a, 0x02, 0x8c, 0x2a, 0x3a, 0x2b, 0xc9, 0x47,
|
||||
0x61, 0x02, 0xbb, 0x28, 0x82, 0x34, 0xc4, 0x15, 0xa2,
|
||||
0xb0, 0x1f, 0x82, 0x8e, 0xa6, 0x2a, 0xc5, 0xb3, 0xe4,
|
||||
0x2f
|
||||
};
|
||||
EXPORT_SYMBOL_GPL(sha224_zero_message_hash);
|
||||
|
||||
const u8 sha256_zero_message_hash[SHA256_DIGEST_SIZE] = {
|
||||
0xe3, 0xb0, 0xc4, 0x42, 0x98, 0xfc, 0x1c, 0x14,
|
||||
0x9a, 0xfb, 0xf4, 0xc8, 0x99, 0x6f, 0xb9, 0x24,
|
||||
0x27, 0xae, 0x41, 0xe4, 0x64, 0x9b, 0x93, 0x4c,
|
||||
0xa4, 0x95, 0x99, 0x1b, 0x78, 0x52, 0xb8, 0x55
|
||||
};
|
||||
EXPORT_SYMBOL_GPL(sha256_zero_message_hash);
|
||||
|
||||
static void sha256_block(struct crypto_sha256_state *sctx, const u8 *input,
|
||||
int blocks)
|
||||
{
|
||||
sha256_transform_blocks(sctx, input, blocks);
|
||||
}
|
||||
|
||||
static int crypto_sha256_update(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len)
|
||||
{
|
||||
return sha256_base_do_update_blocks(desc, data, len, sha256_block);
|
||||
}
|
||||
|
||||
static int crypto_sha256_finup(struct shash_desc *desc, const u8 *data,
|
||||
unsigned int len, u8 *hash)
|
||||
{
|
||||
sha256_base_do_finup(desc, data, len, sha256_block);
|
||||
return sha256_base_finish(desc, hash);
|
||||
}
|
||||
|
||||
static struct shash_alg sha256_algs[2] = { {
|
||||
.digestsize = SHA256_DIGEST_SIZE,
|
||||
.init = sha256_base_init,
|
||||
.update = crypto_sha256_update,
|
||||
.finup = crypto_sha256_finup,
|
||||
.descsize = sizeof(struct crypto_sha256_state),
|
||||
.base = {
|
||||
.cra_name = "sha256",
|
||||
.cra_driver_name= "sha256-generic",
|
||||
.cra_priority = 100,
|
||||
.cra_flags = CRYPTO_AHASH_ALG_BLOCK_ONLY |
|
||||
CRYPTO_AHASH_ALG_FINUP_MAX,
|
||||
.cra_blocksize = SHA256_BLOCK_SIZE,
|
||||
.cra_module = THIS_MODULE,
|
||||
}
|
||||
}, {
|
||||
.digestsize = SHA224_DIGEST_SIZE,
|
||||
.init = sha224_base_init,
|
||||
.update = crypto_sha256_update,
|
||||
.finup = crypto_sha256_finup,
|
||||
.descsize = sizeof(struct crypto_sha256_state),
|
||||
.base = {
|
||||
.cra_name = "sha224",
|
||||
.cra_driver_name= "sha224-generic",
|
||||
.cra_priority = 100,
|
||||
.cra_flags = CRYPTO_AHASH_ALG_BLOCK_ONLY |
|
||||
CRYPTO_AHASH_ALG_FINUP_MAX,
|
||||
.cra_blocksize = SHA224_BLOCK_SIZE,
|
||||
.cra_module = THIS_MODULE,
|
||||
}
|
||||
} };
|
||||
|
||||
static int __init sha256_generic_mod_init(void)
|
||||
{
|
||||
return crypto_register_shashes(sha256_algs, ARRAY_SIZE(sha256_algs));
|
||||
}
|
||||
|
||||
static void __exit sha256_generic_mod_fini(void)
|
||||
{
|
||||
crypto_unregister_shashes(sha256_algs, ARRAY_SIZE(sha256_algs));
|
||||
}
|
||||
|
||||
subsys_initcall(sha256_generic_mod_init);
|
||||
module_exit(sha256_generic_mod_fini);
|
||||
|
||||
MODULE_LICENSE("GPL");
|
||||
MODULE_DESCRIPTION("SHA-224 and SHA-256 Secure Hash Algorithm");
|
||||
|
||||
MODULE_ALIAS_CRYPTO("sha224");
|
||||
MODULE_ALIAS_CRYPTO("sha224-generic");
|
||||
MODULE_ALIAS_CRYPTO("sha256");
|
||||
MODULE_ALIAS_CRYPTO("sha256-generic");
|
28
include/crypto/internal/sha2.h
Normal file
28
include/crypto/internal/sha2.h
Normal file
|
@ -0,0 +1,28 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0-only */
|
||||
|
||||
#ifndef _CRYPTO_INTERNAL_SHA2_H
|
||||
#define _CRYPTO_INTERNAL_SHA2_H
|
||||
|
||||
#include <crypto/sha2.h>
|
||||
|
||||
void sha256_update_generic(struct sha256_state *sctx,
|
||||
const u8 *data, size_t len);
|
||||
void sha256_final_generic(struct sha256_state *sctx,
|
||||
u8 out[SHA256_DIGEST_SIZE]);
|
||||
void sha224_final_generic(struct sha256_state *sctx,
|
||||
u8 out[SHA224_DIGEST_SIZE]);
|
||||
|
||||
#if IS_ENABLED(CONFIG_CRYPTO_ARCH_HAVE_LIB_SHA256)
|
||||
bool sha256_is_arch_optimized(void);
|
||||
#else
|
||||
static inline bool sha256_is_arch_optimized(void)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
#endif
|
||||
void sha256_blocks_generic(u32 state[SHA256_STATE_WORDS],
|
||||
const u8 *data, size_t nblocks);
|
||||
void sha256_blocks_arch(u32 state[SHA256_STATE_WORDS],
|
||||
const u8 *data, size_t nblocks);
|
||||
|
||||
#endif /* _CRYPTO_INTERNAL_SHA2_H */
|
|
@ -13,6 +13,7 @@
|
|||
|
||||
#define SHA256_DIGEST_SIZE 32
|
||||
#define SHA256_BLOCK_SIZE 64
|
||||
#define SHA256_STATE_WORDS 8
|
||||
|
||||
#define SHA384_DIGEST_SIZE 48
|
||||
#define SHA384_BLOCK_SIZE 128
|
||||
|
@ -66,7 +67,7 @@ extern const u8 sha384_zero_message_hash[SHA384_DIGEST_SIZE];
|
|||
extern const u8 sha512_zero_message_hash[SHA512_DIGEST_SIZE];
|
||||
|
||||
struct crypto_sha256_state {
|
||||
u32 state[SHA256_DIGEST_SIZE / 4];
|
||||
u32 state[SHA256_STATE_WORDS];
|
||||
u64 count;
|
||||
};
|
||||
|
||||
|
@ -74,7 +75,7 @@ struct sha256_state {
|
|||
union {
|
||||
struct crypto_sha256_state ctx;
|
||||
struct {
|
||||
u32 state[SHA256_DIGEST_SIZE / 4];
|
||||
u32 state[SHA256_STATE_WORDS];
|
||||
u64 count;
|
||||
};
|
||||
};
|
||||
|
@ -87,16 +88,6 @@ struct sha512_state {
|
|||
u8 buf[SHA512_BLOCK_SIZE];
|
||||
};
|
||||
|
||||
/*
|
||||
* Stand-alone implementation of the SHA256 algorithm. It is designed to
|
||||
* have as little dependencies as possible so it can be used in the
|
||||
* kexec_file purgatory. In other cases you should generally use the
|
||||
* hash APIs from include/crypto/hash.h. Especially when hashing large
|
||||
* amounts of data as those APIs may be hw-accelerated.
|
||||
*
|
||||
* For details see lib/crypto/sha256.c
|
||||
*/
|
||||
|
||||
static inline void sha256_init(struct sha256_state *sctx)
|
||||
{
|
||||
sctx->state[0] = SHA256_H0;
|
||||
|
|
|
@ -10,7 +10,7 @@
|
|||
|
||||
#include <crypto/internal/blockhash.h>
|
||||
#include <crypto/internal/hash.h>
|
||||
#include <crypto/sha2.h>
|
||||
#include <crypto/internal/sha2.h>
|
||||
#include <linux/math.h>
|
||||
#include <linux/string.h>
|
||||
#include <linux/types.h>
|
||||
|
@ -142,7 +142,10 @@ static inline int sha256_base_finish(struct shash_desc *desc, u8 *out)
|
|||
return __sha256_base_finish(sctx->state, out, digest_size);
|
||||
}
|
||||
|
||||
void sha256_transform_blocks(struct crypto_sha256_state *sst,
|
||||
const u8 *input, int blocks);
|
||||
static inline void sha256_transform_blocks(struct crypto_sha256_state *sst,
|
||||
const u8 *input, int blocks)
|
||||
{
|
||||
sha256_blocks_generic(sst->state, input, blocks);
|
||||
}
|
||||
|
||||
#endif /* _CRYPTO_SHA256_BASE_H */
|
||||
|
|
|
@ -139,6 +139,25 @@ config CRYPTO_LIB_SHA1
|
|||
|
||||
config CRYPTO_LIB_SHA256
|
||||
tristate
|
||||
help
|
||||
Enable the SHA-256 library interface. This interface may be fulfilled
|
||||
by either the generic implementation or an arch-specific one, if one
|
||||
is available and enabled.
|
||||
|
||||
config CRYPTO_ARCH_HAVE_LIB_SHA256
|
||||
bool
|
||||
help
|
||||
Declares whether the architecture provides an arch-specific
|
||||
accelerated implementation of the SHA-256 library interface.
|
||||
|
||||
config CRYPTO_LIB_SHA256_GENERIC
|
||||
tristate
|
||||
default CRYPTO_LIB_SHA256 if !CRYPTO_ARCH_HAVE_LIB_SHA256
|
||||
help
|
||||
This symbol can be selected by arch implementations of the SHA-256
|
||||
library interface that require the generic code as a fallback, e.g.,
|
||||
for SIMD implementations. If no arch specific implementation is
|
||||
enabled, this implementation serves the users of CRYPTO_LIB_SHA256.
|
||||
|
||||
config CRYPTO_LIB_SM3
|
||||
tristate
|
||||
|
|
|
@ -51,6 +51,9 @@ libsha1-y := sha1.o
|
|||
obj-$(CONFIG_CRYPTO_LIB_SHA256) += libsha256.o
|
||||
libsha256-y := sha256.o
|
||||
|
||||
obj-$(CONFIG_CRYPTO_LIB_SHA256_GENERIC) += libsha256-generic.o
|
||||
libsha256-generic-y := sha256-generic.o
|
||||
|
||||
ifneq ($(CONFIG_CRYPTO_MANAGER_DISABLE_TESTS),y)
|
||||
libblake2s-y += blake2s-selftest.o
|
||||
libchacha20poly1305-y += chacha20poly1305-selftest.o
|
||||
|
|
137
lib/crypto/sha256-generic.c
Normal file
137
lib/crypto/sha256-generic.c
Normal file
|
@ -0,0 +1,137 @@
|
|||
// SPDX-License-Identifier: GPL-2.0-or-later
|
||||
/*
|
||||
* SHA-256, as specified in
|
||||
* http://csrc.nist.gov/groups/STM/cavp/documents/shs/sha256-384-512.pdf
|
||||
*
|
||||
* SHA-256 code by Jean-Luc Cooke <jlcooke@certainkey.com>.
|
||||
*
|
||||
* Copyright (c) Jean-Luc Cooke <jlcooke@certainkey.com>
|
||||
* Copyright (c) Andrew McDonald <andrew@mcdonald.org.uk>
|
||||
* Copyright (c) 2002 James Morris <jmorris@intercode.com.au>
|
||||
* Copyright (c) 2014 Red Hat Inc.
|
||||
*/
|
||||
|
||||
#include <crypto/internal/sha2.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/string.h>
|
||||
#include <linux/unaligned.h>
|
||||
|
||||
static const u32 SHA256_K[] = {
|
||||
0x428a2f98, 0x71374491, 0xb5c0fbcf, 0xe9b5dba5,
|
||||
0x3956c25b, 0x59f111f1, 0x923f82a4, 0xab1c5ed5,
|
||||
0xd807aa98, 0x12835b01, 0x243185be, 0x550c7dc3,
|
||||
0x72be5d74, 0x80deb1fe, 0x9bdc06a7, 0xc19bf174,
|
||||
0xe49b69c1, 0xefbe4786, 0x0fc19dc6, 0x240ca1cc,
|
||||
0x2de92c6f, 0x4a7484aa, 0x5cb0a9dc, 0x76f988da,
|
||||
0x983e5152, 0xa831c66d, 0xb00327c8, 0xbf597fc7,
|
||||
0xc6e00bf3, 0xd5a79147, 0x06ca6351, 0x14292967,
|
||||
0x27b70a85, 0x2e1b2138, 0x4d2c6dfc, 0x53380d13,
|
||||
0x650a7354, 0x766a0abb, 0x81c2c92e, 0x92722c85,
|
||||
0xa2bfe8a1, 0xa81a664b, 0xc24b8b70, 0xc76c51a3,
|
||||
0xd192e819, 0xd6990624, 0xf40e3585, 0x106aa070,
|
||||
0x19a4c116, 0x1e376c08, 0x2748774c, 0x34b0bcb5,
|
||||
0x391c0cb3, 0x4ed8aa4a, 0x5b9cca4f, 0x682e6ff3,
|
||||
0x748f82ee, 0x78a5636f, 0x84c87814, 0x8cc70208,
|
||||
0x90befffa, 0xa4506ceb, 0xbef9a3f7, 0xc67178f2,
|
||||
};
|
||||
|
||||
static inline u32 Ch(u32 x, u32 y, u32 z)
|
||||
{
|
||||
return z ^ (x & (y ^ z));
|
||||
}
|
||||
|
||||
static inline u32 Maj(u32 x, u32 y, u32 z)
|
||||
{
|
||||
return (x & y) | (z & (x | y));
|
||||
}
|
||||
|
||||
#define e0(x) (ror32(x, 2) ^ ror32(x, 13) ^ ror32(x, 22))
|
||||
#define e1(x) (ror32(x, 6) ^ ror32(x, 11) ^ ror32(x, 25))
|
||||
#define s0(x) (ror32(x, 7) ^ ror32(x, 18) ^ (x >> 3))
|
||||
#define s1(x) (ror32(x, 17) ^ ror32(x, 19) ^ (x >> 10))
|
||||
|
||||
static inline void LOAD_OP(int I, u32 *W, const u8 *input)
|
||||
{
|
||||
W[I] = get_unaligned_be32((__u32 *)input + I);
|
||||
}
|
||||
|
||||
static inline void BLEND_OP(int I, u32 *W)
|
||||
{
|
||||
W[I] = s1(W[I-2]) + W[I-7] + s0(W[I-15]) + W[I-16];
|
||||
}
|
||||
|
||||
#define SHA256_ROUND(i, a, b, c, d, e, f, g, h) do { \
|
||||
u32 t1, t2; \
|
||||
t1 = h + e1(e) + Ch(e, f, g) + SHA256_K[i] + W[i]; \
|
||||
t2 = e0(a) + Maj(a, b, c); \
|
||||
d += t1; \
|
||||
h = t1 + t2; \
|
||||
} while (0)
|
||||
|
||||
static void sha256_block_generic(u32 state[SHA256_STATE_WORDS],
|
||||
const u8 *input, u32 W[64])
|
||||
{
|
||||
u32 a, b, c, d, e, f, g, h;
|
||||
int i;
|
||||
|
||||
/* load the input */
|
||||
for (i = 0; i < 16; i += 8) {
|
||||
LOAD_OP(i + 0, W, input);
|
||||
LOAD_OP(i + 1, W, input);
|
||||
LOAD_OP(i + 2, W, input);
|
||||
LOAD_OP(i + 3, W, input);
|
||||
LOAD_OP(i + 4, W, input);
|
||||
LOAD_OP(i + 5, W, input);
|
||||
LOAD_OP(i + 6, W, input);
|
||||
LOAD_OP(i + 7, W, input);
|
||||
}
|
||||
|
||||
/* now blend */
|
||||
for (i = 16; i < 64; i += 8) {
|
||||
BLEND_OP(i + 0, W);
|
||||
BLEND_OP(i + 1, W);
|
||||
BLEND_OP(i + 2, W);
|
||||
BLEND_OP(i + 3, W);
|
||||
BLEND_OP(i + 4, W);
|
||||
BLEND_OP(i + 5, W);
|
||||
BLEND_OP(i + 6, W);
|
||||
BLEND_OP(i + 7, W);
|
||||
}
|
||||
|
||||
/* load the state into our registers */
|
||||
a = state[0]; b = state[1]; c = state[2]; d = state[3];
|
||||
e = state[4]; f = state[5]; g = state[6]; h = state[7];
|
||||
|
||||
/* now iterate */
|
||||
for (i = 0; i < 64; i += 8) {
|
||||
SHA256_ROUND(i + 0, a, b, c, d, e, f, g, h);
|
||||
SHA256_ROUND(i + 1, h, a, b, c, d, e, f, g);
|
||||
SHA256_ROUND(i + 2, g, h, a, b, c, d, e, f);
|
||||
SHA256_ROUND(i + 3, f, g, h, a, b, c, d, e);
|
||||
SHA256_ROUND(i + 4, e, f, g, h, a, b, c, d);
|
||||
SHA256_ROUND(i + 5, d, e, f, g, h, a, b, c);
|
||||
SHA256_ROUND(i + 6, c, d, e, f, g, h, a, b);
|
||||
SHA256_ROUND(i + 7, b, c, d, e, f, g, h, a);
|
||||
}
|
||||
|
||||
state[0] += a; state[1] += b; state[2] += c; state[3] += d;
|
||||
state[4] += e; state[5] += f; state[6] += g; state[7] += h;
|
||||
}
|
||||
|
||||
void sha256_blocks_generic(u32 state[SHA256_STATE_WORDS],
|
||||
const u8 *data, size_t nblocks)
|
||||
{
|
||||
u32 W[64];
|
||||
|
||||
do {
|
||||
sha256_block_generic(state, data, W);
|
||||
data += SHA256_BLOCK_SIZE;
|
||||
} while (--nblocks);
|
||||
|
||||
memzero_explicit(W, sizeof(W));
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(sha256_blocks_generic);
|
||||
|
||||
MODULE_DESCRIPTION("SHA-256 Algorithm (generic implementation)");
|
||||
MODULE_LICENSE("GPL");
|
|
@ -11,148 +11,105 @@
|
|||
* Copyright (c) 2014 Red Hat Inc.
|
||||
*/
|
||||
|
||||
#include <linux/unaligned.h>
|
||||
#include <crypto/sha256_base.h>
|
||||
#include <crypto/internal/sha2.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/module.h>
|
||||
#include <linux/string.h>
|
||||
#include <linux/unaligned.h>
|
||||
|
||||
static const u32 SHA256_K[] = {
|
||||
0x428a2f98, 0x71374491, 0xb5c0fbcf, 0xe9b5dba5,
|
||||
0x3956c25b, 0x59f111f1, 0x923f82a4, 0xab1c5ed5,
|
||||
0xd807aa98, 0x12835b01, 0x243185be, 0x550c7dc3,
|
||||
0x72be5d74, 0x80deb1fe, 0x9bdc06a7, 0xc19bf174,
|
||||
0xe49b69c1, 0xefbe4786, 0x0fc19dc6, 0x240ca1cc,
|
||||
0x2de92c6f, 0x4a7484aa, 0x5cb0a9dc, 0x76f988da,
|
||||
0x983e5152, 0xa831c66d, 0xb00327c8, 0xbf597fc7,
|
||||
0xc6e00bf3, 0xd5a79147, 0x06ca6351, 0x14292967,
|
||||
0x27b70a85, 0x2e1b2138, 0x4d2c6dfc, 0x53380d13,
|
||||
0x650a7354, 0x766a0abb, 0x81c2c92e, 0x92722c85,
|
||||
0xa2bfe8a1, 0xa81a664b, 0xc24b8b70, 0xc76c51a3,
|
||||
0xd192e819, 0xd6990624, 0xf40e3585, 0x106aa070,
|
||||
0x19a4c116, 0x1e376c08, 0x2748774c, 0x34b0bcb5,
|
||||
0x391c0cb3, 0x4ed8aa4a, 0x5b9cca4f, 0x682e6ff3,
|
||||
0x748f82ee, 0x78a5636f, 0x84c87814, 0x8cc70208,
|
||||
0x90befffa, 0xa4506ceb, 0xbef9a3f7, 0xc67178f2,
|
||||
};
|
||||
/*
|
||||
* If __DISABLE_EXPORTS is defined, then this file is being compiled for a
|
||||
* pre-boot environment. In that case, ignore the kconfig options, pull the
|
||||
* generic code into the same translation unit, and use that only.
|
||||
*/
|
||||
#ifdef __DISABLE_EXPORTS
|
||||
#include "sha256-generic.c"
|
||||
#endif
|
||||
|
||||
static inline u32 Ch(u32 x, u32 y, u32 z)
|
||||
static inline void sha256_blocks(u32 state[SHA256_STATE_WORDS], const u8 *data,
|
||||
size_t nblocks, bool force_generic)
|
||||
{
|
||||
return z ^ (x & (y ^ z));
|
||||
#if IS_ENABLED(CONFIG_CRYPTO_ARCH_HAVE_LIB_SHA256) && !defined(__DISABLE_EXPORTS)
|
||||
if (!force_generic)
|
||||
return sha256_blocks_arch(state, data, nblocks);
|
||||
#endif
|
||||
sha256_blocks_generic(state, data, nblocks);
|
||||
}
|
||||
|
||||
static inline u32 Maj(u32 x, u32 y, u32 z)
|
||||
static inline void __sha256_update(struct sha256_state *sctx, const u8 *data,
|
||||
size_t len, bool force_generic)
|
||||
{
|
||||
return (x & y) | (z & (x | y));
|
||||
}
|
||||
size_t partial = sctx->count % SHA256_BLOCK_SIZE;
|
||||
|
||||
#define e0(x) (ror32(x, 2) ^ ror32(x, 13) ^ ror32(x, 22))
|
||||
#define e1(x) (ror32(x, 6) ^ ror32(x, 11) ^ ror32(x, 25))
|
||||
#define s0(x) (ror32(x, 7) ^ ror32(x, 18) ^ (x >> 3))
|
||||
#define s1(x) (ror32(x, 17) ^ ror32(x, 19) ^ (x >> 10))
|
||||
sctx->count += len;
|
||||
|
||||
static inline void LOAD_OP(int I, u32 *W, const u8 *input)
|
||||
{
|
||||
W[I] = get_unaligned_be32((__u32 *)input + I);
|
||||
}
|
||||
if (partial + len >= SHA256_BLOCK_SIZE) {
|
||||
size_t nblocks;
|
||||
|
||||
static inline void BLEND_OP(int I, u32 *W)
|
||||
{
|
||||
W[I] = s1(W[I-2]) + W[I-7] + s0(W[I-15]) + W[I-16];
|
||||
}
|
||||
if (partial) {
|
||||
size_t l = SHA256_BLOCK_SIZE - partial;
|
||||
|
||||
#define SHA256_ROUND(i, a, b, c, d, e, f, g, h) do { \
|
||||
u32 t1, t2; \
|
||||
t1 = h + e1(e) + Ch(e, f, g) + SHA256_K[i] + W[i]; \
|
||||
t2 = e0(a) + Maj(a, b, c); \
|
||||
d += t1; \
|
||||
h = t1 + t2; \
|
||||
} while (0)
|
||||
memcpy(&sctx->buf[partial], data, l);
|
||||
data += l;
|
||||
len -= l;
|
||||
|
||||
static void sha256_transform(u32 *state, const u8 *input, u32 *W)
|
||||
{
|
||||
u32 a, b, c, d, e, f, g, h;
|
||||
int i;
|
||||
sha256_blocks(sctx->state, sctx->buf, 1, force_generic);
|
||||
}
|
||||
|
||||
/* load the input */
|
||||
for (i = 0; i < 16; i += 8) {
|
||||
LOAD_OP(i + 0, W, input);
|
||||
LOAD_OP(i + 1, W, input);
|
||||
LOAD_OP(i + 2, W, input);
|
||||
LOAD_OP(i + 3, W, input);
|
||||
LOAD_OP(i + 4, W, input);
|
||||
LOAD_OP(i + 5, W, input);
|
||||
LOAD_OP(i + 6, W, input);
|
||||
LOAD_OP(i + 7, W, input);
|
||||
nblocks = len / SHA256_BLOCK_SIZE;
|
||||
len %= SHA256_BLOCK_SIZE;
|
||||
|
||||
if (nblocks) {
|
||||
sha256_blocks(sctx->state, data, nblocks,
|
||||
force_generic);
|
||||
data += nblocks * SHA256_BLOCK_SIZE;
|
||||
}
|
||||
partial = 0;
|
||||
}
|
||||
|
||||
/* now blend */
|
||||
for (i = 16; i < 64; i += 8) {
|
||||
BLEND_OP(i + 0, W);
|
||||
BLEND_OP(i + 1, W);
|
||||
BLEND_OP(i + 2, W);
|
||||
BLEND_OP(i + 3, W);
|
||||
BLEND_OP(i + 4, W);
|
||||
BLEND_OP(i + 5, W);
|
||||
BLEND_OP(i + 6, W);
|
||||
BLEND_OP(i + 7, W);
|
||||
}
|
||||
|
||||
/* load the state into our registers */
|
||||
a = state[0]; b = state[1]; c = state[2]; d = state[3];
|
||||
e = state[4]; f = state[5]; g = state[6]; h = state[7];
|
||||
|
||||
/* now iterate */
|
||||
for (i = 0; i < 64; i += 8) {
|
||||
SHA256_ROUND(i + 0, a, b, c, d, e, f, g, h);
|
||||
SHA256_ROUND(i + 1, h, a, b, c, d, e, f, g);
|
||||
SHA256_ROUND(i + 2, g, h, a, b, c, d, e, f);
|
||||
SHA256_ROUND(i + 3, f, g, h, a, b, c, d, e);
|
||||
SHA256_ROUND(i + 4, e, f, g, h, a, b, c, d);
|
||||
SHA256_ROUND(i + 5, d, e, f, g, h, a, b, c);
|
||||
SHA256_ROUND(i + 6, c, d, e, f, g, h, a, b);
|
||||
SHA256_ROUND(i + 7, b, c, d, e, f, g, h, a);
|
||||
}
|
||||
|
||||
state[0] += a; state[1] += b; state[2] += c; state[3] += d;
|
||||
state[4] += e; state[5] += f; state[6] += g; state[7] += h;
|
||||
if (len)
|
||||
memcpy(&sctx->buf[partial], data, len);
|
||||
}
|
||||
|
||||
void sha256_transform_blocks(struct crypto_sha256_state *sst,
|
||||
const u8 *input, int blocks)
|
||||
{
|
||||
u32 W[64];
|
||||
|
||||
do {
|
||||
sha256_transform(sst->state, input, W);
|
||||
input += SHA256_BLOCK_SIZE;
|
||||
} while (--blocks);
|
||||
|
||||
memzero_explicit(W, sizeof(W));
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(sha256_transform_blocks);
|
||||
|
||||
void sha256_update(struct sha256_state *sctx, const u8 *data, unsigned int len)
|
||||
{
|
||||
lib_sha256_base_do_update(sctx, data, len, sha256_transform_blocks);
|
||||
__sha256_update(sctx, data, len, false);
|
||||
}
|
||||
EXPORT_SYMBOL(sha256_update);
|
||||
|
||||
static void __sha256_final(struct sha256_state *sctx, u8 *out, int digest_size)
|
||||
static inline void __sha256_final(struct sha256_state *sctx, u8 *out,
|
||||
size_t digest_size, bool force_generic)
|
||||
{
|
||||
lib_sha256_base_do_finalize(sctx, sha256_transform_blocks);
|
||||
lib_sha256_base_finish(sctx, out, digest_size);
|
||||
const size_t bit_offset = SHA256_BLOCK_SIZE - sizeof(__be64);
|
||||
__be64 *bits = (__be64 *)&sctx->buf[bit_offset];
|
||||
size_t partial = sctx->count % SHA256_BLOCK_SIZE;
|
||||
size_t i;
|
||||
|
||||
sctx->buf[partial++] = 0x80;
|
||||
if (partial > bit_offset) {
|
||||
memset(&sctx->buf[partial], 0, SHA256_BLOCK_SIZE - partial);
|
||||
sha256_blocks(sctx->state, sctx->buf, 1, force_generic);
|
||||
partial = 0;
|
||||
}
|
||||
|
||||
memset(&sctx->buf[partial], 0, bit_offset - partial);
|
||||
*bits = cpu_to_be64(sctx->count << 3);
|
||||
sha256_blocks(sctx->state, sctx->buf, 1, force_generic);
|
||||
|
||||
for (i = 0; i < digest_size; i += 4)
|
||||
put_unaligned_be32(sctx->state[i / 4], out + i);
|
||||
|
||||
memzero_explicit(sctx, sizeof(*sctx));
|
||||
}
|
||||
|
||||
void sha256_final(struct sha256_state *sctx, u8 *out)
|
||||
{
|
||||
__sha256_final(sctx, out, 32);
|
||||
__sha256_final(sctx, out, SHA256_DIGEST_SIZE, false);
|
||||
}
|
||||
EXPORT_SYMBOL(sha256_final);
|
||||
|
||||
void sha224_final(struct sha256_state *sctx, u8 *out)
|
||||
{
|
||||
__sha256_final(sctx, out, 28);
|
||||
__sha256_final(sctx, out, SHA224_DIGEST_SIZE, false);
|
||||
}
|
||||
EXPORT_SYMBOL(sha224_final);
|
||||
|
||||
|
@ -166,5 +123,26 @@ void sha256(const u8 *data, unsigned int len, u8 *out)
|
|||
}
|
||||
EXPORT_SYMBOL(sha256);
|
||||
|
||||
#if IS_ENABLED(CONFIG_CRYPTO_SHA256) && !defined(__DISABLE_EXPORTS)
|
||||
void sha256_update_generic(struct sha256_state *sctx,
|
||||
const u8 *data, size_t len)
|
||||
{
|
||||
__sha256_update(sctx, data, len, true);
|
||||
}
|
||||
EXPORT_SYMBOL(sha256_update_generic);
|
||||
|
||||
void sha256_final_generic(struct sha256_state *sctx, u8 out[SHA256_DIGEST_SIZE])
|
||||
{
|
||||
__sha256_final(sctx, out, SHA256_DIGEST_SIZE, true);
|
||||
}
|
||||
EXPORT_SYMBOL(sha256_final_generic);
|
||||
|
||||
void sha224_final_generic(struct sha256_state *sctx, u8 out[SHA224_DIGEST_SIZE])
|
||||
{
|
||||
__sha256_final(sctx, out, SHA224_DIGEST_SIZE, true);
|
||||
}
|
||||
EXPORT_SYMBOL(sha224_final_generic);
|
||||
#endif
|
||||
|
||||
MODULE_DESCRIPTION("SHA-256 Algorithm");
|
||||
MODULE_LICENSE("GPL");
|
||||
|
|
Loading…
Add table
Reference in a new issue