//! Low-level "hazmat" AES functions: AES-NI support. //! //! Note: this isn't actually used in the `Aes128`/`Aes192`/`Aes256` //! implementations in this crate, but instead provides raw AES-NI accelerated //! access to the AES round function gated under the `hazmat` crate feature. use super::{ arch::*, utils::{load8, store8}, }; use crate::{Block, ParBlocks}; /// AES cipher (encrypt) round function. #[allow(clippy::cast_ptr_alignment)] #[target_feature(enable = "aes")] pub(crate) unsafe fn cipher_round(block: &mut Block, round_key: &Block) { // Safety: `loadu` and `storeu` support unaligned access let b = _mm_loadu_si128(block.as_ptr() as *const __m128i); let k = _mm_loadu_si128(round_key.as_ptr() as *const __m128i); let out = _mm_aesenc_si128(b, k); _mm_storeu_si128(block.as_mut_ptr() as *mut __m128i, out); } /// AES cipher (encrypt) round function: parallel version. #[allow(clippy::cast_ptr_alignment)] #[target_feature(enable = "aes")] pub(crate) unsafe fn cipher_round_par(blocks: &mut ParBlocks, round_keys: &ParBlocks) { let xmm_keys = load8(round_keys); let mut xmm_blocks = load8(blocks); for i in 0..8 { xmm_blocks[i] = _mm_aesenc_si128(xmm_blocks[i], xmm_keys[i]); } store8(blocks, xmm_blocks); } /// AES cipher (encrypt) round function. #[allow(clippy::cast_ptr_alignment)] #[target_feature(enable = "aes")] pub(crate) unsafe fn equiv_inv_cipher_round(block: &mut Block, round_key: &Block) { // Safety: `loadu` and `storeu` support unaligned access let b = _mm_loadu_si128(block.as_ptr() as *const __m128i); let k = _mm_loadu_si128(round_key.as_ptr() as *const __m128i); let out = _mm_aesdec_si128(b, k); _mm_storeu_si128(block.as_mut_ptr() as *mut __m128i, out); } /// AES cipher (encrypt) round function: parallel version. #[allow(clippy::cast_ptr_alignment)] #[target_feature(enable = "aes")] pub(crate) unsafe fn equiv_inv_cipher_round_par(blocks: &mut ParBlocks, round_keys: &ParBlocks) { let xmm_keys = load8(round_keys); let mut xmm_blocks = load8(blocks); for i in 0..8 { xmm_blocks[i] = _mm_aesdec_si128(xmm_blocks[i], xmm_keys[i]); } store8(blocks, xmm_blocks); } /// AES mix columns function. #[allow(clippy::cast_ptr_alignment)] #[target_feature(enable = "aes")] pub(crate) unsafe fn mix_columns(block: &mut Block) { // Safety: `loadu` and `storeu` support unaligned access let mut state = _mm_loadu_si128(block.as_ptr() as *const __m128i); // Emulate mix columns by performing three inverse mix columns operations state = _mm_aesimc_si128(state); state = _mm_aesimc_si128(state); state = _mm_aesimc_si128(state); _mm_storeu_si128(block.as_mut_ptr() as *mut __m128i, state); } /// AES inverse mix columns function. #[allow(clippy::cast_ptr_alignment)] #[target_feature(enable = "aes")] pub(crate) unsafe fn inv_mix_columns(block: &mut Block) { // Safety: `loadu` and `storeu` support unaligned access let b = _mm_loadu_si128(block.as_ptr() as *const __m128i); let out = _mm_aesimc_si128(b); _mm_storeu_si128(block.as_mut_ptr() as *mut __m128i, out); }