| 1 | // Copyright 2018-2024 Brian Smith. |
| 2 | // |
| 3 | // Permission to use, copy, modify, and/or distribute this software for any |
| 4 | // purpose with or without fee is hereby granted, provided that the above |
| 5 | // copyright notice and this permission notice appear in all copies. |
| 6 | // |
| 7 | // THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHORS DISCLAIM ALL WARRANTIES |
| 8 | // WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF |
| 9 | // MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY |
| 10 | // SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES |
| 11 | // WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN ACTION |
| 12 | // OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF OR IN |
| 13 | // CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE. |
| 14 | |
| 15 | #![cfg (any( |
| 16 | all(target_arch = "aarch64" , target_endian = "little" ), |
| 17 | all(target_arch = "arm" , target_endian = "little" ), |
| 18 | target_arch = "x86" , |
| 19 | target_arch = "x86_64" |
| 20 | ))] |
| 21 | |
| 22 | use super::{Block, Counter, EncryptBlock, EncryptCtr32, Iv, KeyBytes, Overlapping, AES_KEY}; |
| 23 | use crate::{cpu, error}; |
| 24 | |
| 25 | #[cfg (any( |
| 26 | all(target_arch = "aarch64" , target_endian = "little" ), |
| 27 | all(target_arch = "arm" , target_endian = "little" ) |
| 28 | ))] |
| 29 | type RequiredCpuFeatures = cpu::arm::Neon; |
| 30 | |
| 31 | #[cfg (any(target_arch = "x86" , target_arch = "x86_64" ))] |
| 32 | pub(in super::super) type RequiredCpuFeatures = cpu::intel::Ssse3; |
| 33 | |
| 34 | #[derive (Clone)] |
| 35 | pub(in super::super) struct Key { |
| 36 | inner: AES_KEY, |
| 37 | } |
| 38 | |
| 39 | impl Key { |
| 40 | pub(in super::super) fn new( |
| 41 | bytes: KeyBytes<'_>, |
| 42 | _cpu: RequiredCpuFeatures, |
| 43 | ) -> Result<Self, error::Unspecified> { |
| 44 | let inner: AES_KEY = unsafe { set_encrypt_key!(vpaes_set_encrypt_key, bytes) }?; |
| 45 | Ok(Self { inner }) |
| 46 | } |
| 47 | } |
| 48 | |
| 49 | #[cfg (any( |
| 50 | all(target_arch = "aarch64" , target_endian = "little" ), |
| 51 | all(target_arch = "arm" , target_endian = "little" ), |
| 52 | target_arch = "x86_64" |
| 53 | ))] |
| 54 | impl EncryptBlock for Key { |
| 55 | fn encrypt_block(&self, block: Block) -> Block { |
| 56 | super::encrypt_block_using_encrypt_iv_xor_block(self, block) |
| 57 | } |
| 58 | |
| 59 | fn encrypt_iv_xor_block(&self, iv: Iv, block: Block) -> Block { |
| 60 | super::encrypt_iv_xor_block_using_ctr32(self, iv, block) |
| 61 | } |
| 62 | } |
| 63 | |
| 64 | #[cfg (any( |
| 65 | all(target_arch = "aarch64" , target_endian = "little" ), |
| 66 | target_arch = "x86_64" |
| 67 | ))] |
| 68 | impl EncryptCtr32 for Key { |
| 69 | fn ctr32_encrypt_within(&self, in_out: Overlapping<'_>, ctr: &mut Counter) { |
| 70 | unsafe { ctr32_encrypt_blocks!(vpaes_ctr32_encrypt_blocks, in_out, &self.inner, ctr) } |
| 71 | } |
| 72 | } |
| 73 | |
| 74 | #[cfg (all(target_arch = "arm" , target_endian = "little" ))] |
| 75 | impl EncryptCtr32 for Key { |
| 76 | fn ctr32_encrypt_within(&self, in_out: Overlapping<'_>, ctr: &mut Counter) { |
| 77 | use super::{super::overlapping::IndexError, bs, BLOCK_LEN}; |
| 78 | |
| 79 | let in_out = { |
| 80 | let (in_out, src) = in_out.into_slice_src_mut(); |
| 81 | let blocks = in_out[src.clone()].len() / BLOCK_LEN; |
| 82 | |
| 83 | // bsaes operates in batches of 8 blocks. |
| 84 | let bsaes_blocks = if blocks >= 8 && (blocks % 8) < 6 { |
| 85 | // It's faster to use bsaes for all the full batches and then |
| 86 | // switch to vpaes for the last partial batch (if any). |
| 87 | blocks - (blocks % 8) |
| 88 | } else if blocks >= 8 { |
| 89 | // It's faster to let bsaes handle everything including |
| 90 | // the last partial batch. |
| 91 | blocks |
| 92 | } else { |
| 93 | // It's faster to let vpaes handle everything. |
| 94 | 0 |
| 95 | }; |
| 96 | let bsaes_in_out_len = bsaes_blocks * BLOCK_LEN; |
| 97 | let bs_in_out = |
| 98 | Overlapping::new(&mut in_out[..(src.start + bsaes_in_out_len)], src.clone()) |
| 99 | .unwrap_or_else(|IndexError { .. }| unreachable!()); |
| 100 | |
| 101 | // SAFETY: |
| 102 | // * self.inner was initialized with `vpaes_set_encrypt_key` above, |
| 103 | // as required by `bsaes_ctr32_encrypt_blocks_with_vpaes_key`. |
| 104 | unsafe { |
| 105 | bs::ctr32_encrypt_blocks_with_vpaes_key(bs_in_out, &self.inner, ctr); |
| 106 | } |
| 107 | |
| 108 | Overlapping::new(&mut in_out[bsaes_in_out_len..], src) |
| 109 | .unwrap_or_else(|IndexError { .. }| unreachable!()) |
| 110 | }; |
| 111 | |
| 112 | // SAFETY: |
| 113 | // * self.inner was initialized with `vpaes_set_encrypt_key` above, |
| 114 | // as required by `vpaes_ctr32_encrypt_blocks`. |
| 115 | // * `vpaes_ctr32_encrypt_blocks` satisfies the contract for |
| 116 | // `ctr32_encrypt_blocks`. |
| 117 | unsafe { ctr32_encrypt_blocks!(vpaes_ctr32_encrypt_blocks, in_out, &self.inner, ctr) } |
| 118 | } |
| 119 | } |
| 120 | |
| 121 | #[cfg (target_arch = "x86" )] |
| 122 | impl EncryptBlock for Key { |
| 123 | fn encrypt_block(&self, block: Block) -> Block { |
| 124 | unsafe { encrypt_block!(vpaes_encrypt, block, &self.inner) } |
| 125 | } |
| 126 | |
| 127 | fn encrypt_iv_xor_block(&self, iv: Iv, block: Block) -> Block { |
| 128 | super::encrypt_iv_xor_block_using_encrypt_block(self, iv, block) |
| 129 | } |
| 130 | } |
| 131 | |
| 132 | #[cfg (target_arch = "x86" )] |
| 133 | impl EncryptCtr32 for Key { |
| 134 | fn ctr32_encrypt_within(&self, in_out: Overlapping<'_>, ctr: &mut Counter) { |
| 135 | super::super::shift::shift_full_blocks(in_out, |input| { |
| 136 | self.encrypt_iv_xor_block(ctr.increment(), *input) |
| 137 | }); |
| 138 | } |
| 139 | } |
| 140 | |