use super::*;
use crate::WALLET_KEYS_SEED_LEN;
use crate::logger::log_error;
use crate::peer_store::PeerStore;
use crate::{Error, EventQueue, PaymentDetails};
use lightning::chain::channelmonitor::ChannelMonitor;
use lightning::chain::keysinterface::{EntropySource, SignerProvider};
use lightning::routing::gossip::NetworkGraph;
use lightning::routing::scoring::{ProbabilisticScorer, ProbabilisticScoringParameters};
use lightning::util::logger::Logger;
use lightning::util::ser::{Readable, ReadableArgs, Writeable};
use bip39::Mnemonic;
use bitcoin::hash_types::{BlockHash, Txid};
use bitcoin::hashes::hex::FromHex;
use rand::{thread_rng, RngCore};
use std::fs;
use std::io::Write;
use std::ops::Deref;
use std::path::Path;
use std::sync::Arc;
use super::KVStore;
pub fn generate_entropy_mnemonic() -> Mnemonic {
let mut entropy = [0; 32];
thread_rng().fill_bytes(&mut entropy);
Mnemonic::from_entropy(&entropy).unwrap()
}
pub(crate) fn read_or_generate_seed_file<L: Deref>(
keys_seed_path: &str, logger: L,
) -> std::io::Result<[u8; WALLET_KEYS_SEED_LEN]>
where
L::Target: Logger,
{
if Path::new(&keys_seed_path).exists() {
let seed = fs::read(keys_seed_path).map_err(|e| {
log_error!(logger, "Failed to read keys seed file: {}", keys_seed_path);
e
})?;
if seed.len() != WALLET_KEYS_SEED_LEN {
log_error!(
logger,
"Failed to read keys seed file due to invalid length: {}",
keys_seed_path
);
return Err(std::io::Error::new(
std::io::ErrorKind::InvalidData,
"Failed to read keys seed file due to invalid length",
));
}
let mut key = [0; WALLET_KEYS_SEED_LEN];
key.copy_from_slice(&seed);
Ok(key)
} else {
let mut key = [0; WALLET_KEYS_SEED_LEN];
thread_rng().fill_bytes(&mut key);
let mut f = fs::File::create(keys_seed_path).map_err(|e| {
log_error!(logger, "Failed to create keys seed file: {}", keys_seed_path);
e
})?;
f.write_all(&key).map_err(|e| {
log_error!(logger, "Failed to write node keys seed to disk: {}", keys_seed_path);
e
})?;
f.sync_all().map_err(|e| {
log_error!(logger, "Failed to sync node keys seed to disk: {}", keys_seed_path);
e
})?;
Ok(key)
}
}
pub(crate) fn read_channel_monitors<K: KVStore + Sync + Send, ES: Deref, SP: Deref>(
kv_store: Arc<K>, entropy_source: ES, signer_provider: SP,
) -> std::io::Result<Vec<(BlockHash, ChannelMonitor<<SP::Target as SignerProvider>::Signer>)>>
where
ES::Target: EntropySource + Sized,
SP::Target: SignerProvider + Sized,
{
let mut res = Vec::new();
for stored_key in kv_store.list(CHANNEL_MONITOR_PERSISTENCE_NAMESPACE)? {
let txid = Txid::from_hex(stored_key.split_at(64).0).map_err(|_| {
std::io::Error::new(std::io::ErrorKind::InvalidData, "Invalid tx ID in stored key")
})?;
let index: u16 = stored_key.split_at(65).1.parse().map_err(|_| {
std::io::Error::new(std::io::ErrorKind::InvalidData, "Invalid tx index in stored key")
})?;
match <(BlockHash, ChannelMonitor<<SP::Target as SignerProvider>::Signer>)>::read(
&mut kv_store.read(CHANNEL_MONITOR_PERSISTENCE_NAMESPACE, &stored_key)?,
(&*entropy_source, &*signer_provider),
) {
Ok((block_hash, channel_monitor)) => {
if channel_monitor.get_funding_txo().0.txid != txid
|| channel_monitor.get_funding_txo().0.index != index
{
return Err(std::io::Error::new(
std::io::ErrorKind::InvalidData,
"ChannelMonitor was stored under the wrong key",
));
}
res.push((block_hash, channel_monitor));
}
Err(e) => {
return Err(std::io::Error::new(
std::io::ErrorKind::InvalidData,
format!("Failed to deserialize ChannelMonitor: {}", e),
))
}
}
}
Ok(res)
}
pub(crate) fn read_network_graph<K: KVStore + Sync + Send, L: Deref + Clone>(
kv_store: Arc<K>, logger: L,
) -> Result<NetworkGraph<L>, std::io::Error>
where
L::Target: Logger,
{
let mut reader =
kv_store.read(NETWORK_GRAPH_PERSISTENCE_NAMESPACE, NETWORK_GRAPH_PERSISTENCE_KEY)?;
NetworkGraph::read(&mut reader, logger.clone()).map_err(|e| {
log_error!(logger, "Failed to deserialize NetworkGraph: {}", e);
std::io::Error::new(std::io::ErrorKind::InvalidData, "Failed to deserialize NetworkGraph")
})
}
pub(crate) fn read_scorer<
K: KVStore + Send + Sync,
G: Deref<Target = NetworkGraph<L>>,
L: Deref + Clone,
>(
kv_store: Arc<K>, network_graph: G, logger: L,
) -> Result<ProbabilisticScorer<G, L>, std::io::Error>
where
L::Target: Logger,
{
let params = ProbabilisticScoringParameters::default();
let mut reader = kv_store.read(SCORER_PERSISTENCE_NAMESPACE, SCORER_PERSISTENCE_KEY)?;
let args = (params, network_graph, logger.clone());
ProbabilisticScorer::read(&mut reader, args).map_err(|e| {
log_error!(logger, "Failed to deserialize scorer: {}", e);
std::io::Error::new(std::io::ErrorKind::InvalidData, "Failed to deserialize Scorer")
})
}
pub(crate) fn read_event_queue<K: KVStore + Sync + Send, L: Deref + Clone>(
kv_store: Arc<K>, logger: L,
) -> Result<EventQueue<K, L>, std::io::Error>
where
L::Target: Logger,
{
let mut reader =
kv_store.read(EVENT_QUEUE_PERSISTENCE_NAMESPACE, EVENT_QUEUE_PERSISTENCE_KEY)?;
EventQueue::read(&mut reader, (kv_store, logger.clone())).map_err(|e| {
log_error!(logger, "Failed to deserialize event queue: {}", e);
std::io::Error::new(std::io::ErrorKind::InvalidData, "Failed to deserialize EventQueue")
})
}
pub(crate) fn read_peer_info<K: KVStore + Sync + Send, L: Deref + Clone>(
kv_store: Arc<K>, logger: L,
) -> Result<PeerStore<K, L>, std::io::Error>
where
L::Target: Logger,
{
let mut reader = kv_store.read(PEER_INFO_PERSISTENCE_NAMESPACE, PEER_INFO_PERSISTENCE_KEY)?;
PeerStore::read(&mut reader, (kv_store, logger.clone())).map_err(|e| {
log_error!(logger, "Failed to deserialize peer store: {}", e);
std::io::Error::new(std::io::ErrorKind::InvalidData, "Failed to deserialize PeerStore")
})
}
pub(crate) fn read_payments<K: KVStore + Sync + Send, L: Deref>(
kv_store: Arc<K>, logger: L,
) -> Result<Vec<PaymentDetails>, std::io::Error>
where
L::Target: Logger,
{
let mut res = Vec::new();
for stored_key in kv_store.list(PAYMENT_INFO_PERSISTENCE_NAMESPACE)? {
let payment = PaymentDetails::read(
&mut kv_store.read(PAYMENT_INFO_PERSISTENCE_NAMESPACE, &stored_key)?,
)
.map_err(|e| {
log_error!(logger, "Failed to deserialize PaymentDetails: {}", e);
std::io::Error::new(
std::io::ErrorKind::InvalidData,
"Failed to deserialize PaymentDetails",
)
})?;
res.push(payment);
}
Ok(res)
}
pub(crate) fn read_latest_rgs_sync_timestamp<K: KVStore + Sync + Send, L: Deref>(
kv_store: Arc<K>, logger: L,
) -> Result<u32, std::io::Error>
where
L::Target: Logger,
{
let mut reader =
kv_store.read(LATEST_RGS_SYNC_TIMESTAMP_NAMESPACE, LATEST_RGS_SYNC_TIMESTAMP_KEY)?;
u32::read(&mut reader).map_err(|e| {
log_error!(logger, "Failed to deserialize latest RGS sync timestamp: {}", e);
std::io::Error::new(
std::io::ErrorKind::InvalidData,
"Failed to deserialize latest RGS sync timestamp",
)
})
}
pub(crate) fn write_latest_rgs_sync_timestamp<K: KVStore + Sync + Send, L: Deref>(
updated_timestamp: u32, kv_store: Arc<K>, logger: L,
) -> Result<(), Error>
where
L::Target: Logger,
{
let data = updated_timestamp.encode();
kv_store
.write(LATEST_RGS_SYNC_TIMESTAMP_NAMESPACE, LATEST_RGS_SYNC_TIMESTAMP_KEY, &data)
.map_err(|e| {
log_error!(
logger,
"Writing data to key {}/{} failed due to: {}",
LATEST_RGS_SYNC_TIMESTAMP_NAMESPACE,
LATEST_RGS_SYNC_TIMESTAMP_KEY,
e
);
Error::PersistenceFailed
})
}
pub(crate) fn read_latest_node_ann_bcast_timestamp<K: KVStore + Sync + Send, L: Deref>(
kv_store: Arc<K>, logger: L,
) -> Result<u64, std::io::Error>
where
L::Target: Logger,
{
let mut reader = kv_store
.read(LATEST_NODE_ANN_BCAST_TIMESTAMP_NAMESPACE, LATEST_NODE_ANN_BCAST_TIMESTAMP_KEY)?;
u64::read(&mut reader).map_err(|e| {
log_error!(
logger,
"Failed to deserialize latest node announcement broadcast timestamp: {}",
e
);
std::io::Error::new(
std::io::ErrorKind::InvalidData,
"Failed to deserialize latest node announcement broadcast timestamp",
)
})
}
pub(crate) fn write_latest_node_ann_bcast_timestamp<K: KVStore + Sync + Send, L: Deref>(
updated_timestamp: u64, kv_store: Arc<K>, logger: L,
) -> Result<(), Error>
where
L::Target: Logger,
{
let data = updated_timestamp.encode();
kv_store
.write(
LATEST_NODE_ANN_BCAST_TIMESTAMP_NAMESPACE,
LATEST_NODE_ANN_BCAST_TIMESTAMP_KEY,
&data,
)
.map_err(|e| {
log_error!(
logger,
"Writing data to key {}/{} failed due to: {}",
LATEST_NODE_ANN_BCAST_TIMESTAMP_NAMESPACE,
LATEST_NODE_ANN_BCAST_TIMESTAMP_KEY,
e
);
Error::PersistenceFailed
})
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn mnemonic_to_entropy_to_mnemonic() {
let mnemonic = generate_entropy_mnemonic();
let entropy = mnemonic.to_entropy();
assert_eq!(mnemonic, Mnemonic::from_entropy(&entropy).unwrap());
}
}