use std::path::Path;
use std::sync::atomic::{AtomicU64, Ordering};
use std::sync::Arc;
use bytes::Bytes;
use exoware_server::StoreEngine;
use rocksdb::{Direction, IteratorMode, Options, DB};
const SEQ_META_KEY: &[u8] = b"__simulator_seq__";
#[derive(Clone)]
pub struct RocksStore {
db: Arc<DB>,
sequence: Arc<AtomicU64>,
observer: Option<Arc<AtomicU64>>,
}
impl RocksStore {
pub fn open(path: &Path) -> Result<Self, rocksdb::Error> {
Self::open_with_observer(path, None)
}
pub fn open_with_observer(
path: &Path,
observer: Option<Arc<AtomicU64>>,
) -> Result<Self, rocksdb::Error> {
let mut opts = Options::default();
opts.create_if_missing(true);
let db = Arc::new(DB::open(&opts, path)?);
let seq = match db.get(SEQ_META_KEY)? {
Some(bytes) if bytes.len() == 8 => u64::from_le_bytes(bytes.try_into().unwrap()),
_ => 0,
};
Ok(Self {
db,
sequence: Arc::new(AtomicU64::new(seq)),
observer,
})
}
fn batch_put_rocksdb(&self, kvs: &[(Bytes, Bytes)]) -> Result<u64, rocksdb::Error> {
let next = self.sequence.fetch_add(1, Ordering::SeqCst) + 1;
let mut batch = rocksdb::WriteBatch::default();
for (k, v) in kvs {
batch.put(k.as_ref(), v.as_ref());
}
batch.put(SEQ_META_KEY, next.to_le_bytes());
self.db.write(batch)?;
if let Some(obs) = &self.observer {
obs.store(next, Ordering::SeqCst);
}
Ok(next)
}
fn get_rocksdb(&self, key: &[u8]) -> Result<Option<Vec<u8>>, rocksdb::Error> {
if key == SEQ_META_KEY {
return Ok(None);
}
self.db.get(key)
}
fn range_scan_rocksdb(
&self,
start: &[u8],
end: &[u8],
limit: usize,
forward: bool,
) -> Result<Vec<(Bytes, Bytes)>, rocksdb::Error> {
if limit == 0 {
return Ok(Vec::new());
}
let mode = IteratorMode::From(start, Direction::Forward);
let mut tmp = Vec::new();
for item in self.db.iterator(mode) {
let (k, v) = item?;
if k.as_ref() == SEQ_META_KEY {
continue;
}
if k.as_ref() < start {
continue;
}
if !end.is_empty() && k.as_ref() > end {
break;
}
tmp.push((
Bytes::copy_from_slice(k.as_ref()),
Bytes::copy_from_slice(&v),
));
}
if tmp.is_empty() {
return Ok(tmp);
}
if forward {
tmp.truncate(limit);
return Ok(tmp);
}
if tmp.len() > limit {
tmp = tmp.split_off(tmp.len() - limit);
}
tmp.reverse();
Ok(tmp)
}
}
impl StoreEngine for RocksStore {
fn put_batch(&self, kvs: &[(Bytes, Bytes)]) -> Result<u64, String> {
self.batch_put_rocksdb(kvs).map_err(|e| e.to_string())
}
fn get(&self, key: &[u8]) -> Result<Option<Vec<u8>>, String> {
self.get_rocksdb(key).map_err(|e| e.to_string())
}
fn range_scan(
&self,
start: &[u8],
end: &[u8],
limit: usize,
forward: bool,
) -> Result<Vec<(Bytes, Bytes)>, String> {
self.range_scan_rocksdb(start, end, limit, forward)
.map_err(|e| e.to_string())
}
fn get_many(&self, keys: &[&[u8]]) -> Result<Vec<(Vec<u8>, Option<Vec<u8>>)>, String> {
let results = self.db.multi_get(keys);
keys.iter()
.zip(results)
.map(|(k, r)| {
if *k == SEQ_META_KEY {
return Ok((k.to_vec(), None));
}
let value = r.map_err(|e| e.to_string())?;
Ok((k.to_vec(), value))
})
.collect()
}
fn delete_batch(&self, keys: &[&[u8]]) -> Result<u64, String> {
let next = self.sequence.fetch_add(1, Ordering::SeqCst) + 1;
let mut batch = rocksdb::WriteBatch::default();
for k in keys {
batch.delete(k);
}
batch.put(SEQ_META_KEY, next.to_le_bytes());
self.db.write(batch).map_err(|e| e.to_string())?;
if let Some(obs) = &self.observer {
obs.store(next, Ordering::SeqCst);
}
Ok(next)
}
fn current_sequence(&self) -> u64 {
self.sequence.load(Ordering::SeqCst)
}
}