mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-08-05 16:54:27 +00:00

Restore the SIMD usability check that was removed by commita59e5468a9
("crypto: arm64/poly1305 - Add block-only interface"). This safety check is cheap and is well worth eliminating a footgun. While the Poly1305 functions should not be called when SIMD registers are unusable, if they are anyway, they should just do the right thing instead of corrupting random tasks' registers and/or computing incorrect MACs. Fixing this is also needed for poly1305_kunit to pass. Just use may_use_simd() instead of the original crypto_simd_usable(), since poly1305_kunit won't rely on crypto_simd_disabled_for_test. Fixes:a59e5468a9
("crypto: arm64/poly1305 - Add block-only interface") Cc: stable@vger.kernel.org Reviewed-by: Ard Biesheuvel <ardb@kernel.org> Link: https://lore.kernel.org/r/20250706231100.176113-4-ebiggers@kernel.org Signed-off-by: Eric Biggers <ebiggers@kernel.org>
74 lines
2.1 KiB
C
74 lines
2.1 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* OpenSSL/Cryptogams accelerated Poly1305 transform for arm64
|
|
*
|
|
* Copyright (C) 2019 Linaro Ltd. <ard.biesheuvel@linaro.org>
|
|
*/
|
|
|
|
#include <asm/hwcap.h>
|
|
#include <asm/neon.h>
|
|
#include <asm/simd.h>
|
|
#include <crypto/internal/poly1305.h>
|
|
#include <linux/cpufeature.h>
|
|
#include <linux/jump_label.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/unaligned.h>
|
|
|
|
asmlinkage void poly1305_block_init_arch(
|
|
struct poly1305_block_state *state,
|
|
const u8 raw_key[POLY1305_BLOCK_SIZE]);
|
|
EXPORT_SYMBOL_GPL(poly1305_block_init_arch);
|
|
asmlinkage void poly1305_blocks(struct poly1305_block_state *state,
|
|
const u8 *src, u32 len, u32 hibit);
|
|
asmlinkage void poly1305_blocks_neon(struct poly1305_block_state *state,
|
|
const u8 *src, u32 len, u32 hibit);
|
|
asmlinkage void poly1305_emit_arch(const struct poly1305_state *state,
|
|
u8 digest[POLY1305_DIGEST_SIZE],
|
|
const u32 nonce[4]);
|
|
EXPORT_SYMBOL_GPL(poly1305_emit_arch);
|
|
|
|
static __ro_after_init DEFINE_STATIC_KEY_FALSE(have_neon);
|
|
|
|
void poly1305_blocks_arch(struct poly1305_block_state *state, const u8 *src,
|
|
unsigned int len, u32 padbit)
|
|
{
|
|
len = round_down(len, POLY1305_BLOCK_SIZE);
|
|
if (static_branch_likely(&have_neon) && likely(may_use_simd())) {
|
|
do {
|
|
unsigned int todo = min_t(unsigned int, len, SZ_4K);
|
|
|
|
kernel_neon_begin();
|
|
poly1305_blocks_neon(state, src, todo, padbit);
|
|
kernel_neon_end();
|
|
|
|
len -= todo;
|
|
src += todo;
|
|
} while (len);
|
|
} else
|
|
poly1305_blocks(state, src, len, padbit);
|
|
}
|
|
EXPORT_SYMBOL_GPL(poly1305_blocks_arch);
|
|
|
|
bool poly1305_is_arch_optimized(void)
|
|
{
|
|
/* We always can use at least the ARM64 scalar implementation. */
|
|
return true;
|
|
}
|
|
EXPORT_SYMBOL(poly1305_is_arch_optimized);
|
|
|
|
static int __init neon_poly1305_mod_init(void)
|
|
{
|
|
if (cpu_have_named_feature(ASIMD))
|
|
static_branch_enable(&have_neon);
|
|
return 0;
|
|
}
|
|
subsys_initcall(neon_poly1305_mod_init);
|
|
|
|
static void __exit neon_poly1305_mod_exit(void)
|
|
{
|
|
}
|
|
module_exit(neon_poly1305_mod_exit);
|
|
|
|
MODULE_DESCRIPTION("Poly1305 authenticator (ARM64 optimized)");
|
|
MODULE_LICENSE("GPL v2");
|