/* * Carry-less multiply operations. * SPDX-License-Identifier: GPL-2.0-or-later * * Copyright (C) 2023 Linaro, Ltd. */ #ifndef CRYPTO_CLMUL_H #define CRYPTO_CLMUL_H #include "qemu/int128.h" #include "host/crypto/clmul.h" /** * clmul_8x8_low: * * Perform eight 8x8->8 carry-less multiplies. */ uint64_t clmul_8x8_low(uint64_t, uint64_t); /** * clmul_8x4_even: * * Perform four 8x8->16 carry-less multiplies. * The odd bytes of the inputs are ignored. */ uint64_t clmul_8x4_even(uint64_t, uint64_t); /** * clmul_8x4_odd: * * Perform four 8x8->16 carry-less multiplies. * The even bytes of the inputs are ignored. */ uint64_t clmul_8x4_odd(uint64_t, uint64_t); /** * clmul_8x4_packed: * * Perform four 8x8->16 carry-less multiplies. */ uint64_t clmul_8x4_packed(uint32_t, uint32_t); /** * clmul_16x2_even: * * Perform two 16x16->32 carry-less multiplies. * The odd words of the inputs are ignored. */ uint64_t clmul_16x2_even(uint64_t, uint64_t); /** * clmul_16x2_odd: * * Perform two 16x16->32 carry-less multiplies. * The even words of the inputs are ignored. */ uint64_t clmul_16x2_odd(uint64_t, uint64_t); /** * clmul_32: * * Perform a 32x32->64 carry-less multiply. */ uint64_t clmul_32(uint32_t, uint32_t); /** * clmul_64: * * Perform a 64x64->128 carry-less multiply. */ Int128 clmul_64_gen(uint64_t, uint64_t); static inline Int128 clmul_64(uint64_t a, uint64_t b) { if (HAVE_CLMUL_ACCEL) { return clmul_64_accel(a, b); } else { return clmul_64_gen(a, b); } } #endif /* CRYPTO_CLMUL_H */