| 1 | use core::slice; |
| 2 | use core::mem::MaybeUninit; |
| 3 | |
| 4 | use byteorder::{ByteOrder, LE}; |
| 5 | |
| 6 | use crate::Hasher as _; |
| 7 | |
| 8 | /// 32-bit MurmurHash3 hasher |
| 9 | pub struct Hasher { |
| 10 | buf: Buffer, |
| 11 | index: Index, |
| 12 | processed: u32, |
| 13 | state: State, |
| 14 | } |
| 15 | |
| 16 | struct State(u32); |
| 17 | |
| 18 | #[derive (Clone, Copy)] |
| 19 | #[repr (align(4))] |
| 20 | struct Buffer { |
| 21 | bytes: MaybeUninit<[u8; 4]>, |
| 22 | } |
| 23 | |
| 24 | #[derive (Clone, Copy, PartialEq)] |
| 25 | enum Index { |
| 26 | _0, |
| 27 | _1, |
| 28 | _2, |
| 29 | _3, |
| 30 | } |
| 31 | |
| 32 | impl Index { |
| 33 | fn usize(&self) -> usize { |
| 34 | match *self { |
| 35 | Index::_0 => 0, |
| 36 | Index::_1 => 1, |
| 37 | Index::_2 => 2, |
| 38 | Index::_3 => 3, |
| 39 | } |
| 40 | } |
| 41 | } |
| 42 | |
| 43 | impl From<usize> for Index { |
| 44 | fn from(x: usize) -> Self { |
| 45 | match x % 4 { |
| 46 | 0 => Index::_0, |
| 47 | 1 => Index::_1, |
| 48 | 2 => Index::_2, |
| 49 | 3 => Index::_3, |
| 50 | _ => unreachable!(), |
| 51 | } |
| 52 | } |
| 53 | } |
| 54 | |
| 55 | impl Hasher { |
| 56 | fn push(&mut self, buf: &[u8]) { |
| 57 | let start: usize = self.index.usize(); |
| 58 | let len: usize = buf.len(); |
| 59 | // NOTE(unsafe) avoid calling `memcpy` on a 0-3 byte copy |
| 60 | // self.buf.bytes[start..start+len].copy_from(buf); |
| 61 | for i: usize in 0..len { |
| 62 | unsafe { |
| 63 | *self.buf.bytes.assume_init_mut().get_unchecked_mut(index:start + i) = *buf.get_unchecked(index:i); |
| 64 | } |
| 65 | } |
| 66 | self.index = Index::from(start + len); |
| 67 | } |
| 68 | } |
| 69 | |
| 70 | impl Default for Hasher { |
| 71 | #[allow (deprecated)] |
| 72 | fn default() -> Self { |
| 73 | Hasher { |
| 74 | buf: Buffer { bytes: MaybeUninit::uninit() }, |
| 75 | index: Index::_0, |
| 76 | processed: 0, |
| 77 | state: State(0), |
| 78 | } |
| 79 | } |
| 80 | } |
| 81 | |
| 82 | impl crate::Hasher for Hasher { |
| 83 | fn finish32(&self) -> u32 { |
| 84 | // tail |
| 85 | let mut state = match self.index { |
| 86 | Index::_3 => { |
| 87 | let mut block = 0; |
| 88 | unsafe { |
| 89 | block ^= u32::from(self.buf.bytes.assume_init_ref()[2]) << 16; |
| 90 | block ^= u32::from(self.buf.bytes.assume_init_ref()[1]) << 8; |
| 91 | block ^= u32::from(self.buf.bytes.assume_init_ref()[0]); |
| 92 | } |
| 93 | self.state.0 ^ pre_mix(block) |
| 94 | } |
| 95 | Index::_2 => { |
| 96 | let mut block = 0; |
| 97 | unsafe { |
| 98 | block ^= u32::from(self.buf.bytes.assume_init_ref()[1]) << 8; |
| 99 | block ^= u32::from(self.buf.bytes.assume_init_ref()[0]); |
| 100 | } |
| 101 | self.state.0 ^ pre_mix(block) |
| 102 | } |
| 103 | Index::_1 => { |
| 104 | let mut block = 0; |
| 105 | unsafe { |
| 106 | block ^= u32::from(self.buf.bytes.assume_init_ref()[0]); |
| 107 | } |
| 108 | self.state.0 ^ pre_mix(block) |
| 109 | } |
| 110 | Index::_0 => self.state.0, |
| 111 | }; |
| 112 | |
| 113 | // finalization mix |
| 114 | state ^= self.processed; |
| 115 | state ^= state >> 16; |
| 116 | state = state.wrapping_mul(0x85ebca6b); |
| 117 | state ^= state >> 13; |
| 118 | state = state.wrapping_mul(0xc2b2ae35); |
| 119 | state ^= state >> 16; |
| 120 | |
| 121 | state |
| 122 | } |
| 123 | } |
| 124 | |
| 125 | impl core::hash::Hasher for Hasher { |
| 126 | #[inline ] |
| 127 | fn write(&mut self, bytes: &[u8]) { |
| 128 | let len = bytes.len(); |
| 129 | self.processed += len as u32; |
| 130 | |
| 131 | let body = if self.index == Index::_0 { |
| 132 | bytes |
| 133 | } else { |
| 134 | let index = self.index.usize(); |
| 135 | if len + index >= 4 { |
| 136 | // we can complete a block using the data left in the buffer |
| 137 | // NOTE(unsafe) avoid panicking branch (`slice_index_len_fail`) |
| 138 | // let (head, body) = bytes.split_at(4 - index); |
| 139 | let mid = 4 - index; |
| 140 | let head = unsafe { slice::from_raw_parts(bytes.as_ptr(), mid) }; |
| 141 | let body = unsafe { |
| 142 | slice::from_raw_parts(bytes.as_ptr().offset(mid as isize), len - mid) |
| 143 | }; |
| 144 | |
| 145 | // NOTE(unsafe) avoid calling `memcpy` on a 0-3 byte copy |
| 146 | // self.buf.bytes[index..].copy_from_slice(head); |
| 147 | for i in 0..4 - index { |
| 148 | unsafe { |
| 149 | *self.buf.bytes.assume_init_mut().get_unchecked_mut(index + i) = *head.get_unchecked(i); |
| 150 | } |
| 151 | } |
| 152 | |
| 153 | self.index = Index::_0; |
| 154 | |
| 155 | self.state.process_block(&self.buf.bytes); |
| 156 | |
| 157 | body |
| 158 | } else { |
| 159 | bytes |
| 160 | } |
| 161 | }; |
| 162 | |
| 163 | for block in body.chunks(4) { |
| 164 | if block.len() == 4 { |
| 165 | self.state |
| 166 | .process_block(unsafe { &*(block.as_ptr() as *const _) }); |
| 167 | } else { |
| 168 | self.push(block); |
| 169 | } |
| 170 | } |
| 171 | |
| 172 | // XXX is this faster? |
| 173 | // for block in body.exact_chunks(4) { |
| 174 | // self.state |
| 175 | // .process_block(unsafe { &*(block.as_ptr() as *const _) }); |
| 176 | // } |
| 177 | |
| 178 | // let tail = body.split_at(body.len() / 4 * 4).1; |
| 179 | |
| 180 | // self.push(tail); |
| 181 | } |
| 182 | |
| 183 | #[inline ] |
| 184 | fn finish(&self) -> u64 { |
| 185 | self.finish32().into() |
| 186 | } |
| 187 | } |
| 188 | |
| 189 | const C1: u32 = 0xcc9e2d51; |
| 190 | const C2: u32 = 0x1b873593; |
| 191 | const R1: u32 = 15; |
| 192 | |
| 193 | impl State { |
| 194 | fn process_block(&mut self, block: &MaybeUninit<[u8; 4]>) { |
| 195 | self.0 ^= pre_mix(block:LE::read_u32(buf:unsafe { block.assume_init_ref() })); |
| 196 | self.0 = self.0.rotate_left(13); |
| 197 | self.0 = 5u32.wrapping_mul(self.0).wrapping_add(0xe6546b64); |
| 198 | } |
| 199 | } |
| 200 | |
| 201 | fn pre_mix(mut block: u32) -> u32 { |
| 202 | block = block.wrapping_mul(C1); |
| 203 | block = block.rotate_left(R1); |
| 204 | block = block.wrapping_mul(C2); |
| 205 | block |
| 206 | } |
| 207 | |