use crate::utils::{check_namespace_key_validity, is_valid_kvstore_str};
use lightning::types::string::PrintableString;
use lightning::util::persist::{KVStoreSync, MigratableKVStore};
use std::collections::HashMap;
use std::fs;
use std::io::{Read, Write};
use std::path::{Path, PathBuf};
use std::sync::atomic::{AtomicU64, AtomicUsize, Ordering};
use std::sync::{Arc, Mutex, RwLock};
#[cfg(feature = "tokio")]
use core::future::Future;
#[cfg(feature = "tokio")]
use core::pin::Pin;
#[cfg(feature = "tokio")]
use lightning::util::persist::KVStore;
#[cfg(target_os = "windows")]
use {std::ffi::OsStr, std::os::windows::ffi::OsStrExt};
#[cfg(target_os = "windows")]
macro_rules! call {
($e: expr) => {
if $e != 0 {
Ok(())
} else {
Err(std::io::Error::last_os_error())
}
};
}
#[cfg(target_os = "windows")]
fn path_to_windows_str<T: AsRef<OsStr>>(path: &T) -> Vec<u16> {
path.as_ref().encode_wide().chain(Some(0)).collect()
}
const LIST_DIR_CONSISTENCY_RETRIES: usize = 10;
struct FilesystemStoreInner {
data_dir: PathBuf,
tmp_file_counter: AtomicUsize,
locks: Mutex<HashMap<PathBuf, Arc<RwLock<u64>>>>,
}
pub struct FilesystemStore {
inner: Arc<FilesystemStoreInner>,
next_version: AtomicU64,
}
impl FilesystemStore {
pub fn new(data_dir: PathBuf) -> Self {
let locks = Mutex::new(HashMap::new());
let tmp_file_counter = AtomicUsize::new(0);
Self {
inner: Arc::new(FilesystemStoreInner { data_dir, tmp_file_counter, locks }),
next_version: AtomicU64::new(1),
}
}
pub fn get_data_dir(&self) -> PathBuf {
self.inner.data_dir.clone()
}
fn get_new_version_and_lock_ref(&self, dest_file_path: PathBuf) -> (Arc<RwLock<u64>>, u64) {
let version = self.next_version.fetch_add(1, Ordering::Relaxed);
if version == u64::MAX {
panic!("FilesystemStore version counter overflowed");
}
let inner_lock_ref = self.inner.get_inner_lock_ref(dest_file_path);
(inner_lock_ref, version)
}
#[cfg(any(all(feature = "tokio", test), fuzzing))]
pub fn state_size(&self) -> usize {
let outer_lock = self.inner.locks.lock().unwrap();
outer_lock.len()
}
}
impl KVStoreSync for FilesystemStore {
fn read(
&self, primary_namespace: &str, secondary_namespace: &str, key: &str,
) -> Result<Vec<u8>, lightning::io::Error> {
let path = self.inner.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
Some(key),
"read",
)?;
self.inner.read(path)
}
fn write(
&self, primary_namespace: &str, secondary_namespace: &str, key: &str, buf: Vec<u8>,
) -> Result<(), lightning::io::Error> {
let path = self.inner.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
Some(key),
"write",
)?;
let (inner_lock_ref, version) = self.get_new_version_and_lock_ref(path.clone());
self.inner.write_version(inner_lock_ref, path, buf, version)
}
fn remove(
&self, primary_namespace: &str, secondary_namespace: &str, key: &str, lazy: bool,
) -> Result<(), lightning::io::Error> {
let path = self.inner.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
Some(key),
"remove",
)?;
let (inner_lock_ref, version) = self.get_new_version_and_lock_ref(path.clone());
self.inner.remove_version(inner_lock_ref, path, lazy, version)
}
fn list(
&self, primary_namespace: &str, secondary_namespace: &str,
) -> Result<Vec<String>, lightning::io::Error> {
let path = self.inner.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
None,
"list",
)?;
self.inner.list(path)
}
}
impl FilesystemStoreInner {
fn get_inner_lock_ref(&self, path: PathBuf) -> Arc<RwLock<u64>> {
let mut outer_lock = self.locks.lock().unwrap();
Arc::clone(&outer_lock.entry(path).or_default())
}
fn get_dest_dir_path(
&self, primary_namespace: &str, secondary_namespace: &str,
) -> std::io::Result<PathBuf> {
let mut dest_dir_path = {
#[cfg(target_os = "windows")]
{
let data_dir = self.data_dir.clone();
fs::create_dir_all(data_dir.clone())?;
fs::canonicalize(data_dir)?
}
#[cfg(not(target_os = "windows"))]
{
self.data_dir.clone()
}
};
dest_dir_path.push(primary_namespace);
if !secondary_namespace.is_empty() {
dest_dir_path.push(secondary_namespace);
}
Ok(dest_dir_path)
}
fn get_checked_dest_file_path(
&self, primary_namespace: &str, secondary_namespace: &str, key: Option<&str>,
operation: &str,
) -> lightning::io::Result<PathBuf> {
check_namespace_key_validity(primary_namespace, secondary_namespace, key, operation)?;
let mut dest_file_path = self.get_dest_dir_path(primary_namespace, secondary_namespace)?;
if let Some(key) = key {
dest_file_path.push(key);
}
Ok(dest_file_path)
}
fn read(&self, dest_file_path: PathBuf) -> lightning::io::Result<Vec<u8>> {
let mut buf = Vec::new();
self.execute_locked_read(dest_file_path.clone(), || {
let mut f = fs::File::open(dest_file_path)?;
f.read_to_end(&mut buf)?;
Ok(())
})?;
Ok(buf)
}
fn execute_locked_write<F: FnOnce() -> Result<(), lightning::io::Error>>(
&self, inner_lock_ref: Arc<RwLock<u64>>, dest_file_path: PathBuf, version: u64, callback: F,
) -> Result<(), lightning::io::Error> {
let res = {
let mut last_written_version = inner_lock_ref.write().unwrap();
let is_stale_version = version <= *last_written_version;
if is_stale_version {
Ok(())
} else {
callback().map(|_| {
*last_written_version = version;
})
}
};
self.clean_locks(&inner_lock_ref, dest_file_path);
res
}
fn execute_locked_read<F: FnOnce() -> Result<(), lightning::io::Error>>(
&self, dest_file_path: PathBuf, callback: F,
) -> Result<(), lightning::io::Error> {
let inner_lock_ref = self.get_inner_lock_ref(dest_file_path.clone());
let res = {
let _guard = inner_lock_ref.read().unwrap();
callback()
};
self.clean_locks(&inner_lock_ref, dest_file_path);
res
}
fn clean_locks(&self, inner_lock_ref: &Arc<RwLock<u64>>, dest_file_path: PathBuf) {
let mut outer_lock = self.locks.lock().unwrap();
let strong_count = Arc::strong_count(&inner_lock_ref);
debug_assert!(strong_count >= 2, "Unexpected FilesystemStore strong count");
if strong_count == 2 {
outer_lock.remove(&dest_file_path);
}
}
fn write_version(
&self, inner_lock_ref: Arc<RwLock<u64>>, dest_file_path: PathBuf, buf: Vec<u8>,
version: u64,
) -> lightning::io::Result<()> {
let parent_directory = dest_file_path.parent().ok_or_else(|| {
let msg =
format!("Could not retrieve parent directory of {}.", dest_file_path.display());
std::io::Error::new(std::io::ErrorKind::InvalidInput, msg)
})?;
fs::create_dir_all(&parent_directory)?;
let mut tmp_file_path = dest_file_path.clone();
let tmp_file_ext = format!("{}.tmp", self.tmp_file_counter.fetch_add(1, Ordering::AcqRel));
tmp_file_path.set_extension(tmp_file_ext);
{
let mut tmp_file = fs::File::create(&tmp_file_path)?;
tmp_file.write_all(&buf)?;
tmp_file.sync_all()?;
}
self.execute_locked_write(inner_lock_ref, dest_file_path.clone(), version, || {
#[cfg(not(target_os = "windows"))]
{
fs::rename(&tmp_file_path, &dest_file_path)?;
let dir_file = fs::OpenOptions::new().read(true).open(&parent_directory)?;
dir_file.sync_all()?;
Ok(())
}
#[cfg(target_os = "windows")]
{
let res = if dest_file_path.exists() {
call!(unsafe {
windows_sys::Win32::Storage::FileSystem::ReplaceFileW(
path_to_windows_str(&dest_file_path).as_ptr(),
path_to_windows_str(&tmp_file_path).as_ptr(),
std::ptr::null(),
windows_sys::Win32::Storage::FileSystem::REPLACEFILE_IGNORE_MERGE_ERRORS,
std::ptr::null_mut() as *const core::ffi::c_void,
std::ptr::null_mut() as *const core::ffi::c_void,
)
})
} else {
call!(unsafe {
windows_sys::Win32::Storage::FileSystem::MoveFileExW(
path_to_windows_str(&tmp_file_path).as_ptr(),
path_to_windows_str(&dest_file_path).as_ptr(),
windows_sys::Win32::Storage::FileSystem::MOVEFILE_WRITE_THROUGH
| windows_sys::Win32::Storage::FileSystem::MOVEFILE_REPLACE_EXISTING,
)
})
};
match res {
Ok(()) => {
let dest_file =
fs::OpenOptions::new().read(true).write(true).open(&dest_file_path)?;
dest_file.sync_all()?;
Ok(())
},
Err(e) => Err(e.into()),
}
}
})
}
fn remove_version(
&self, inner_lock_ref: Arc<RwLock<u64>>, dest_file_path: PathBuf, lazy: bool, version: u64,
) -> lightning::io::Result<()> {
self.execute_locked_write(inner_lock_ref, dest_file_path.clone(), version, || {
if !dest_file_path.is_file() {
return Ok(());
}
if lazy {
fs::remove_file(&dest_file_path)?;
} else {
#[cfg(not(target_os = "windows"))]
{
fs::remove_file(&dest_file_path)?;
let parent_directory = dest_file_path.parent().ok_or_else(|| {
let msg = format!(
"Could not retrieve parent directory of {}.",
dest_file_path.display()
);
std::io::Error::new(std::io::ErrorKind::InvalidInput, msg)
})?;
let dir_file = fs::OpenOptions::new().read(true).open(parent_directory)?;
dir_file.sync_all()?;
}
#[cfg(target_os = "windows")]
{
let mut trash_file_path = dest_file_path.clone();
let trash_file_ext =
format!("{}.trash", self.tmp_file_counter.fetch_add(1, Ordering::AcqRel));
trash_file_path.set_extension(trash_file_ext);
call!(unsafe {
windows_sys::Win32::Storage::FileSystem::MoveFileExW(
path_to_windows_str(&dest_file_path).as_ptr(),
path_to_windows_str(&trash_file_path).as_ptr(),
windows_sys::Win32::Storage::FileSystem::MOVEFILE_WRITE_THROUGH
| windows_sys::Win32::Storage::FileSystem::MOVEFILE_REPLACE_EXISTING,
)
})?;
{
let trash_file = fs::OpenOptions::new()
.read(true)
.write(true)
.open(&trash_file_path.clone())?;
trash_file.sync_all()?;
}
fs::remove_file(trash_file_path).ok();
}
}
Ok(())
})
}
fn list(&self, prefixed_dest: PathBuf) -> lightning::io::Result<Vec<String>> {
if !Path::new(&prefixed_dest).exists() {
return Ok(Vec::new());
}
let mut keys;
let mut retries = LIST_DIR_CONSISTENCY_RETRIES;
'retry_list: loop {
keys = Vec::new();
'skip_entry: for entry in fs::read_dir(&prefixed_dest)? {
let entry = entry?;
let p = entry.path();
let res = dir_entry_is_key(&entry);
match res {
Ok(true) => {
let key = get_key_from_dir_entry_path(&p, &prefixed_dest)?;
keys.push(key);
},
Ok(false) => {
continue 'skip_entry;
},
Err(e) => {
if e.kind() == lightning::io::ErrorKind::NotFound && retries > 0 {
retries -= 1;
continue 'retry_list;
} else {
return Err(e.into());
}
},
}
}
break 'retry_list;
}
Ok(keys)
}
}
#[cfg(feature = "tokio")]
impl KVStore for FilesystemStore {
fn read(
&self, primary_namespace: &str, secondary_namespace: &str, key: &str,
) -> Pin<Box<dyn Future<Output = Result<Vec<u8>, lightning::io::Error>> + 'static + Send>> {
let this = Arc::clone(&self.inner);
let path = match this.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
Some(key),
"read",
) {
Ok(path) => path,
Err(e) => return Box::pin(async move { Err(e) }),
};
Box::pin(async move {
tokio::task::spawn_blocking(move || this.read(path)).await.unwrap_or_else(|e| {
Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, e))
})
})
}
fn write(
&self, primary_namespace: &str, secondary_namespace: &str, key: &str, buf: Vec<u8>,
) -> Pin<Box<dyn Future<Output = Result<(), lightning::io::Error>> + 'static + Send>> {
let this = Arc::clone(&self.inner);
let path = match this.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
Some(key),
"write",
) {
Ok(path) => path,
Err(e) => return Box::pin(async move { Err(e) }),
};
let (inner_lock_ref, version) = self.get_new_version_and_lock_ref(path.clone());
Box::pin(async move {
tokio::task::spawn_blocking(move || {
this.write_version(inner_lock_ref, path, buf, version)
})
.await
.unwrap_or_else(|e| Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, e)))
})
}
fn remove(
&self, primary_namespace: &str, secondary_namespace: &str, key: &str, lazy: bool,
) -> Pin<Box<dyn Future<Output = Result<(), lightning::io::Error>> + 'static + Send>> {
let this = Arc::clone(&self.inner);
let path = match this.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
Some(key),
"remove",
) {
Ok(path) => path,
Err(e) => return Box::pin(async move { Err(e) }),
};
let (inner_lock_ref, version) = self.get_new_version_and_lock_ref(path.clone());
Box::pin(async move {
tokio::task::spawn_blocking(move || {
this.remove_version(inner_lock_ref, path, lazy, version)
})
.await
.unwrap_or_else(|e| Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, e)))
})
}
fn list(
&self, primary_namespace: &str, secondary_namespace: &str,
) -> Pin<Box<dyn Future<Output = Result<Vec<String>, lightning::io::Error>> + 'static + Send>> {
let this = Arc::clone(&self.inner);
let path = match this.get_checked_dest_file_path(
primary_namespace,
secondary_namespace,
None,
"list",
) {
Ok(path) => path,
Err(e) => return Box::pin(async move { Err(e) }),
};
Box::pin(async move {
tokio::task::spawn_blocking(move || this.list(path)).await.unwrap_or_else(|e| {
Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, e))
})
})
}
}
fn dir_entry_is_key(dir_entry: &fs::DirEntry) -> Result<bool, lightning::io::Error> {
let p = dir_entry.path();
if let Some(ext) = p.extension() {
#[cfg(target_os = "windows")]
{
if ext == "trash" {
fs::remove_file(p).ok();
return Ok(false);
}
}
if ext == "tmp" {
return Ok(false);
}
}
let metadata = dir_entry.metadata()?;
if metadata.is_dir() {
return Ok(false);
}
if !metadata.is_file() {
debug_assert!(
false,
"Failed to list keys at path {}: file couldn't be accessed.",
PrintableString(p.to_str().unwrap_or_default())
);
let msg = format!(
"Failed to list keys at path {}: file couldn't be accessed.",
PrintableString(p.to_str().unwrap_or_default())
);
return Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, msg));
}
Ok(true)
}
fn get_key_from_dir_entry_path(p: &Path, base_path: &Path) -> Result<String, lightning::io::Error> {
match p.strip_prefix(&base_path) {
Ok(stripped_path) => {
if let Some(relative_path) = stripped_path.to_str() {
if is_valid_kvstore_str(relative_path) {
return Ok(relative_path.to_string());
} else {
debug_assert!(
false,
"Failed to list keys of path {}: file path is not valid key",
PrintableString(p.to_str().unwrap_or_default())
);
let msg = format!(
"Failed to list keys of path {}: file path is not valid key",
PrintableString(p.to_str().unwrap_or_default())
);
return Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, msg));
}
} else {
debug_assert!(
false,
"Failed to list keys of path {}: file path is not valid UTF-8",
PrintableString(p.to_str().unwrap_or_default())
);
let msg = format!(
"Failed to list keys of path {}: file path is not valid UTF-8",
PrintableString(p.to_str().unwrap_or_default())
);
return Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, msg));
}
},
Err(e) => {
debug_assert!(
false,
"Failed to list keys of path {}: {}",
PrintableString(p.to_str().unwrap_or_default()),
e
);
let msg = format!(
"Failed to list keys of path {}: {}",
PrintableString(p.to_str().unwrap_or_default()),
e
);
return Err(lightning::io::Error::new(lightning::io::ErrorKind::Other, msg));
},
}
}
impl MigratableKVStore for FilesystemStore {
fn list_all_keys(&self) -> Result<Vec<(String, String, String)>, lightning::io::Error> {
let prefixed_dest = &self.inner.data_dir;
if !prefixed_dest.exists() {
return Ok(Vec::new());
}
let mut keys = Vec::new();
'primary_loop: for primary_entry in fs::read_dir(prefixed_dest)? {
let primary_entry = primary_entry?;
let primary_path = primary_entry.path();
if dir_entry_is_key(&primary_entry)? {
let primary_namespace = String::new();
let secondary_namespace = String::new();
let key = get_key_from_dir_entry_path(&primary_path, prefixed_dest)?;
keys.push((primary_namespace, secondary_namespace, key));
continue 'primary_loop;
}
'secondary_loop: for secondary_entry in fs::read_dir(&primary_path)? {
let secondary_entry = secondary_entry?;
let secondary_path = secondary_entry.path();
if dir_entry_is_key(&secondary_entry)? {
let primary_namespace =
get_key_from_dir_entry_path(&primary_path, prefixed_dest)?;
let secondary_namespace = String::new();
let key = get_key_from_dir_entry_path(&secondary_path, &primary_path)?;
keys.push((primary_namespace, secondary_namespace, key));
continue 'secondary_loop;
}
for tertiary_entry in fs::read_dir(&secondary_path)? {
let tertiary_entry = tertiary_entry?;
let tertiary_path = tertiary_entry.path();
if dir_entry_is_key(&tertiary_entry)? {
let primary_namespace =
get_key_from_dir_entry_path(&primary_path, prefixed_dest)?;
let secondary_namespace =
get_key_from_dir_entry_path(&secondary_path, &primary_path)?;
let key = get_key_from_dir_entry_path(&tertiary_path, &secondary_path)?;
keys.push((primary_namespace, secondary_namespace, key));
} else {
debug_assert!(
false,
"Failed to list keys of path {}: only two levels of namespaces are supported",
PrintableString(tertiary_path.to_str().unwrap_or_default())
);
let msg = format!(
"Failed to list keys of path {}: only two levels of namespaces are supported",
PrintableString(tertiary_path.to_str().unwrap_or_default())
);
return Err(lightning::io::Error::new(
lightning::io::ErrorKind::Other,
msg,
));
}
}
}
}
Ok(keys)
}
}
#[cfg(test)]
mod tests {
use super::*;
use crate::test_utils::{
do_read_write_remove_list_persist, do_test_data_migration, do_test_store,
};
use lightning::chain::chainmonitor::Persist;
use lightning::chain::ChannelMonitorUpdateStatus;
use lightning::check_closed_event;
use lightning::events::ClosureReason;
use lightning::ln::functional_test_utils::*;
use lightning::ln::msgs::BaseMessageHandler;
use lightning::util::persist::read_channel_monitors;
use lightning::util::test_utils;
impl Drop for FilesystemStore {
fn drop(&mut self) {
match fs::remove_dir_all(&self.inner.data_dir) {
Err(e) => println!("Failed to remove test persister directory: {}", e),
_ => {},
}
}
}
#[test]
fn read_write_remove_list_persist() {
let mut temp_path = std::env::temp_dir();
temp_path.push("test_read_write_remove_list_persist");
let fs_store = FilesystemStore::new(temp_path);
do_read_write_remove_list_persist(&fs_store);
}
#[cfg(feature = "tokio")]
#[tokio::test]
async fn read_write_remove_list_persist_async() {
use crate::fs_store::FilesystemStore;
use lightning::util::persist::KVStore;
use std::sync::Arc;
let mut temp_path = std::env::temp_dir();
temp_path.push("test_read_write_remove_list_persist_async");
let fs_store = Arc::new(FilesystemStore::new(temp_path));
assert_eq!(fs_store.state_size(), 0);
let async_fs_store: Arc<dyn KVStore> = fs_store.clone();
let data1 = vec![42u8; 32];
let data2 = vec![43u8; 32];
let primary_namespace = "testspace";
let secondary_namespace = "testsubspace";
let key = "testkey";
let fut1 = async_fs_store.write(primary_namespace, secondary_namespace, key, data1);
assert_eq!(fs_store.state_size(), 1);
let fut2 = async_fs_store.remove(primary_namespace, secondary_namespace, key, false);
assert_eq!(fs_store.state_size(), 1);
let fut3 = async_fs_store.write(primary_namespace, secondary_namespace, key, data2.clone());
assert_eq!(fs_store.state_size(), 1);
fut3.await.unwrap();
assert_eq!(fs_store.state_size(), 1);
fut2.await.unwrap();
assert_eq!(fs_store.state_size(), 1);
fut1.await.unwrap();
assert_eq!(fs_store.state_size(), 0);
let listed_keys =
async_fs_store.list(primary_namespace, secondary_namespace).await.unwrap();
assert_eq!(listed_keys.len(), 1);
assert_eq!(listed_keys[0], key);
let read_data =
async_fs_store.read(primary_namespace, secondary_namespace, key).await.unwrap();
assert_eq!(data2, &*read_data);
async_fs_store.remove(primary_namespace, secondary_namespace, key, false).await.unwrap();
let listed_keys =
async_fs_store.list(primary_namespace, secondary_namespace).await.unwrap();
assert_eq!(listed_keys.len(), 0);
}
#[test]
fn test_data_migration() {
let mut source_temp_path = std::env::temp_dir();
source_temp_path.push("test_data_migration_source");
let mut source_store = FilesystemStore::new(source_temp_path);
let mut target_temp_path = std::env::temp_dir();
target_temp_path.push("test_data_migration_target");
let mut target_store = FilesystemStore::new(target_temp_path);
do_test_data_migration(&mut source_store, &mut target_store);
}
#[test]
fn test_if_monitors_is_not_dir() {
let store = FilesystemStore::new("test_monitors_is_not_dir".into());
fs::create_dir_all(&store.get_data_dir()).unwrap();
let mut path = std::path::PathBuf::from(&store.get_data_dir());
path.push("monitors");
fs::File::create(path).unwrap();
let chanmon_cfgs = create_chanmon_cfgs(1);
let mut node_cfgs = create_node_cfgs(1, &chanmon_cfgs);
let chain_mon_0 = test_utils::TestChainMonitor::new(
Some(&chanmon_cfgs[0].chain_source),
&chanmon_cfgs[0].tx_broadcaster,
&chanmon_cfgs[0].logger,
&chanmon_cfgs[0].fee_estimator,
&store,
node_cfgs[0].keys_manager,
);
node_cfgs[0].chain_monitor = chain_mon_0;
let node_chanmgrs = create_node_chanmgrs(1, &node_cfgs, &[None]);
let nodes = create_network(1, &node_cfgs, &node_chanmgrs);
assert!(
read_channel_monitors(&store, nodes[0].keys_manager, nodes[0].keys_manager).is_err()
);
}
#[test]
fn test_filesystem_store() {
let store_0 = FilesystemStore::new("test_filesystem_store_0".into());
let store_1 = FilesystemStore::new("test_filesystem_store_1".into());
do_test_store(&store_0, &store_1)
}
#[cfg(not(target_os = "windows"))]
#[test]
fn test_readonly_dir_perm_failure() {
let store = FilesystemStore::new("test_readonly_dir_perm_failure".into());
fs::create_dir_all(&store.get_data_dir()).unwrap();
let chanmon_cfgs = create_chanmon_cfgs(2);
let node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
let node_a_id = nodes[0].node.get_our_node_id();
let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
let message = "Channel force-closed".to_owned();
nodes[1]
.node
.force_close_broadcasting_latest_txn(&chan.2, &node_a_id, message.clone())
.unwrap();
let reason =
ClosureReason::HolderForceClosed { broadcasted_latest_txn: Some(true), message };
check_closed_event!(nodes[1], 1, reason, [node_a_id], 100000);
let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
let path = &store.get_data_dir();
let mut perms = fs::metadata(path).unwrap().permissions();
perms.set_readonly(true);
fs::set_permissions(path, perms).unwrap();
let monitor_name = added_monitors[0].1.persistence_key();
match store.persist_new_channel(monitor_name, &added_monitors[0].1) {
ChannelMonitorUpdateStatus::UnrecoverableError => {},
_ => panic!("unexpected result from persisting new channel"),
}
nodes[1].node.get_and_clear_pending_msg_events();
added_monitors.clear();
}
#[cfg(target_os = "windows")]
#[test]
fn test_fail_on_open() {
let chanmon_cfgs = create_chanmon_cfgs(2);
let node_cfgs = create_node_cfgs(2, &chanmon_cfgs);
let node_chanmgrs = create_node_chanmgrs(2, &node_cfgs, &[None, None]);
let nodes = create_network(2, &node_cfgs, &node_chanmgrs);
let node_a_id = nodes[0].node.get_our_node_id();
let chan = create_announced_chan_between_nodes(&nodes, 0, 1);
let message = "Channel force-closed".to_owned();
nodes[1]
.node
.force_close_broadcasting_latest_txn(&chan.2, &node_a_id, message.clone())
.unwrap();
let reason =
ClosureReason::HolderForceClosed { broadcasted_latest_txn: Some(true), message };
check_closed_event!(nodes[1], 1, reason, [node_a_id], 100000);
let mut added_monitors = nodes[1].chain_monitor.added_monitors.lock().unwrap();
let update_map = nodes[1].chain_monitor.latest_monitor_update_id.lock().unwrap();
let update_id = update_map.get(&added_monitors[0].1.channel_id()).unwrap();
let store = FilesystemStore::new(":<>/".into());
let monitor_name = added_monitors[0].1.persistence_key();
match store.persist_new_channel(monitor_name, &added_monitors[0].1) {
ChannelMonitorUpdateStatus::UnrecoverableError => {},
_ => panic!("unexpected result from persisting new channel"),
}
nodes[1].node.get_and_clear_pending_msg_events();
added_monitors.clear();
}
}
#[cfg(ldk_bench)]
pub mod bench {
use criterion::Criterion;
pub fn bench_sends(bench: &mut Criterion) {
let store_a = super::FilesystemStore::new("bench_filesystem_store_a".into());
let store_b = super::FilesystemStore::new("bench_filesystem_store_b".into());
lightning::ln::channelmanager::bench::bench_two_sends(
bench,
"bench_filesystem_persisted_sends",
store_a,
store_b,
);
}
}