1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
use std::mem;
use std::ptr;
use std::sync::Arc;
use std::sync::atomic;
use std::sync::atomic::{AtomicPtr, Ordering};
use std::thread;
use crate::{WeakEpoch, Epochs, Inner, OperationCache, USIZE_MSB};
pub struct WriteHandle<T: OperationCache> {
writers_inner: Option<Arc<AtomicPtr<Inner<T>>>>,
readers_inner: Arc<AtomicPtr<Inner<T>>>,
epochs: Epochs,
last_epochs: Vec<usize>,
ops: Vec<T::Operation>,
}
impl<T: OperationCache> WriteHandle<T> {
pub(crate) fn new(writers_inner: Arc<AtomicPtr<Inner<T>>>, readers_inner: Arc<AtomicPtr<Inner<T>>>, epochs: Epochs) -> Self {
Self {
writers_inner: Some(writers_inner),
readers_inner,
epochs,
last_epochs: Vec::new(),
ops: Vec::new(),
}
}
pub fn write(&mut self, operation: T::Operation) {
self.ops.push(operation)
}
fn wait(&mut self, epochs: &mut Vec<WeakEpoch>) {
let mut start_index = 0;
let mut retry_count = 0;
self.last_epochs.resize(epochs.len(), 0);
'retrying: loop {
for index in start_index..self.last_epochs.len() {
let epoch = match epochs[index].upgrade() {
Some(e) => e,
None => {
epochs.remove(index);
self.last_epochs.remove(index);
start_index = 0;
continue 'retrying
}
};
if self.last_epochs[index] & USIZE_MSB != 0 {
continue
}
let current_epoch = epoch.load(Ordering::Acquire);
if current_epoch == self.last_epochs[index] && current_epoch & USIZE_MSB == 0 && current_epoch != 0 {
start_index = index;
if retry_count < 32 {
retry_count += 1;
} else {
thread::yield_now();
}
continue 'retrying
}
}
break
}
}
pub fn refresh(&mut self) {
let epochs = Arc::clone(&self.epochs);
let mut epochs = epochs.lock().unwrap();
self.wait(&mut epochs);
let w_handle = &mut unsafe {
self.writers_inner.as_ref().unwrap().load(Ordering::Relaxed).as_mut().unwrap()
}.value;
for operation in self.ops.iter().cloned() {
w_handle.apply_operation(operation);
}
let writers_inner = self.writers_inner.as_ref().unwrap().swap(self.readers_inner.load(Ordering::Relaxed), Ordering::Release);
self.readers_inner.store(writers_inner, Ordering::Release);
atomic::fence(Ordering::SeqCst);
for (i, epoch) in epochs.iter().enumerate() {
if let Some(e) = epoch.upgrade() {
self.last_epochs[i] = e.load(Ordering::Acquire);
}
}
let w_handle = &mut unsafe {
self.writers_inner.as_ref().unwrap().load(Ordering::Relaxed).as_mut().unwrap()
}.value;
for operation in self.ops.drain(0..self.ops.len()) {
w_handle.apply_operation(operation)
}
}
pub fn into_inner(mut self) -> T {
let writers_inner = self.writers_inner.take().unwrap();
unsafe { Box::from_raw(writers_inner.swap(ptr::null_mut(), Ordering::Relaxed)) }.value
}
}
impl<T: OperationCache> Drop for WriteHandle<T> {
fn drop(&mut self) {
if self.writers_inner.is_some() {
if !self.ops.is_empty() {
self.refresh();
}
assert!(self.ops.is_empty());
let writers_inner = self.writers_inner.as_ref().unwrap().swap(ptr::null_mut(), Ordering::Relaxed);
mem::drop(unsafe { Box::from_raw(writers_inner) });
}
}
}