axhash_core/hash/
hasher_impl.rs1use super::AxHasher;
2
3use crate::bytes::hash_bytes_core;
4use crate::constants::SECRET;
5use crate::math::{avalanche, folded_multiply};
6
7use core::hash::Hasher;
8
9impl Hasher for AxHasher {
10 #[inline(always)]
11 fn finish(&self) -> u64 {
12 if self.sponge_bits == 0 {
13 avalanche(self.acc)
14 } else {
15 let lo = self.sponge as u64;
16 let hi = (self.sponge >> 64) as u64;
17 avalanche(folded_multiply(lo ^ self.acc, hi ^ SECRET[1]))
18 }
19 }
20
21 #[inline(always)]
22 fn write(&mut self, bytes: &[u8]) {
23 self.flush_sponge();
24 self.acc = hash_bytes_core(bytes, self.acc);
25 }
26
27 #[inline(always)]
28 fn write_u8(&mut self, i: u8) {
29 self.push_num(i, 8);
30 }
31
32 #[inline(always)]
33 fn write_u16(&mut self, i: u16) {
34 self.push_num(i, 16);
35 }
36
37 #[inline(always)]
38 fn write_u32(&mut self, i: u32) {
39 self.push_num(i, 32);
40 }
41
42 #[inline(always)]
43 fn write_u64(&mut self, i: u64) {
44 self.push_num(i, 64);
45 }
46
47 #[inline(always)]
48 fn write_u128(&mut self, i: u128) {
49 self.flush_sponge();
50 let lo = i as u64;
51 let hi = (i >> 64) as u64;
52 self.acc = folded_multiply(lo ^ self.acc, hi ^ SECRET[1]);
53 }
54
55 #[inline(always)]
56 fn write_usize(&mut self, i: usize) {
57 #[cfg(target_pointer_width = "32")]
58 self.write_u32(i as u32);
59 #[cfg(target_pointer_width = "64")]
60 self.write_u64(i as u64);
61 }
62
63 #[inline(always)]
64 fn write_i8(&mut self, i: i8) {
65 self.write_u8(i as u8);
66 }
67
68 #[inline(always)]
69 fn write_i16(&mut self, i: i16) {
70 self.write_u16(i as u16);
71 }
72
73 #[inline(always)]
74 fn write_i32(&mut self, i: i32) {
75 self.write_u32(i as u32);
76 }
77
78 #[inline(always)]
79 fn write_i64(&mut self, i: i64) {
80 self.write_u64(i as u64);
81 }
82
83 #[inline(always)]
84 fn write_i128(&mut self, i: i128) {
85 self.write_u128(i as u128);
86 }
87
88 #[inline(always)]
89 fn write_isize(&mut self, i: isize) {
90 self.write_usize(i as usize);
91 }
92}