1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235
// Copyright (C) 2019-2023 Aleo Systems Inc.
// This file is part of the snarkOS library.
// Licensed under the Apache License, Version 2.0 (the "License");
// you may not use this file except in compliance with the License.
// You may obtain a copy of the License at:
// http://www.apache.org/licenses/LICENSE-2.0
// Unless required by applicable law or agreed to in writing, software
// distributed under the License is distributed on an "AS IS" BASIS,
// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
// See the License for the specific language governing permissions and
// limitations under the License.
mod router;
use crate::traits::NodeInterface;
use snarkos_account::Account;
use snarkos_node_consensus::Consensus;
use snarkos_node_messages::{BlockRequest, Message, NodeType, PuzzleResponse, UnconfirmedSolution};
use snarkos_node_rest::Rest;
use snarkos_node_router::{Heartbeat, Inbound, Outbound, Router, Routing};
use snarkos_node_tcp::{
protocols::{Disconnect, Handshake, OnConnect, Reading, Writing},
P2P,
};
use snarkvm::prelude::{Block, ConsensusStorage, Header, Ledger, Network, ProverSolution};
use anyhow::Result;
use parking_lot::Mutex;
use std::{
net::SocketAddr,
sync::{
atomic::{AtomicBool, Ordering},
Arc,
},
time::Duration,
};
use tokio::task::JoinHandle;
/// A validator is a full node, capable of validating blocks.
#[derive(Clone)]
pub struct Validator<N: Network, C: ConsensusStorage<N>> {
/// The ledger of the node.
ledger: Ledger<N, C>,
/// The consensus module of the node.
consensus: Consensus<N, C>,
/// The router of the node.
router: Router<N>,
/// The REST server of the node.
rest: Option<Rest<N, C, Self>>,
/// The spawned handles.
handles: Arc<Mutex<Vec<JoinHandle<()>>>>,
/// The shutdown signal.
shutdown: Arc<AtomicBool>,
}
impl<N: Network, C: ConsensusStorage<N>> Validator<N, C> {
/// Initializes a new validator node.
pub async fn new(
node_ip: SocketAddr,
rest_ip: Option<SocketAddr>,
account: Account<N>,
trusted_peers: &[SocketAddr],
genesis: Block<N>,
cdn: Option<String>,
dev: Option<u16>,
) -> Result<Self> {
// Initialize the ledger.
let ledger = Ledger::load(genesis, dev)?;
// Initialize the CDN.
if let Some(base_url) = cdn {
// Sync the ledger with the CDN.
if let Err((_, error)) = snarkos_node_cdn::sync_ledger_with_cdn(&base_url, ledger.clone()).await {
crate::helpers::log_clean_error(dev);
return Err(error);
}
}
// Initialize the consensus.
let consensus = Consensus::new(ledger.clone(), dev.is_some())?;
// Initialize the node router.
let router = Router::new(
node_ip,
NodeType::Validator,
account,
trusted_peers,
Self::MAXIMUM_NUMBER_OF_PEERS as u16,
dev.is_some(),
)
.await?;
// Initialize the node.
let mut node = Self {
ledger: ledger.clone(),
consensus: consensus.clone(),
router,
rest: None,
handles: Default::default(),
shutdown: Default::default(),
};
// Initialize the REST server.
if let Some(rest_ip) = rest_ip {
node.rest = Some(Rest::start(rest_ip, Some(consensus), ledger, Arc::new(node.clone()))?);
}
// Initialize the sync pool.
node.initialize_sync()?;
// Initialize the routing.
node.initialize_routing().await;
// Initialize the signal handler.
node.handle_signals();
// Return the node.
Ok(node)
}
/// Returns the ledger.
pub fn ledger(&self) -> &Ledger<N, C> {
&self.ledger
}
/// Returns the REST server.
pub fn rest(&self) -> &Option<Rest<N, C, Self>> {
&self.rest
}
}
#[async_trait]
impl<N: Network, C: ConsensusStorage<N>> NodeInterface<N> for Validator<N, C> {
/// Shuts down the node.
async fn shut_down(&self) {
info!("Shutting down...");
// Shut down the sync pool.
trace!("Shutting down the sync pool...");
self.shutdown.store(true, Ordering::Relaxed);
// Abort the tasks.
trace!("Shutting down the validator...");
self.handles.lock().iter().for_each(|handle| handle.abort());
// Shut down the router.
self.router.shut_down().await;
// Shut down the ledger.
trace!("Shutting down the ledger...");
// self.ledger.shut_down().await;
info!("Node has shut down.");
}
}
impl<N: Network, C: ConsensusStorage<N>> Validator<N, C> {
/// Initializes the sync pool.
fn initialize_sync(&self) -> Result<()> {
// Retrieve the canon locators.
let canon_locators = crate::helpers::get_block_locators(&self.ledger)?;
// Insert the canon locators into the sync pool.
self.router.sync().insert_canon_locators(canon_locators).unwrap();
// Start the sync loop.
let validator = self.clone();
self.handles.lock().push(tokio::spawn(async move {
loop {
// If the Ctrl-C handler registered the signal, stop the node.
if validator.shutdown.load(Ordering::Relaxed) {
info!("Shutting down block production");
break;
}
// Sleep briefly to avoid triggering spam detection.
tokio::time::sleep(Duration::from_secs(1)).await;
// Prepare the block requests, if any.
let block_requests = validator.router.sync().prepare_block_requests();
trace!("Prepared {} block requests", block_requests.len());
// Process the block requests.
'outer: for (height, (hash, previous_hash, sync_ips)) in block_requests {
// Insert the block request into the sync pool.
let result =
validator.router.sync().insert_block_request(height, (hash, previous_hash, sync_ips.clone()));
// If the block request was inserted, send it to the peers.
if result.is_ok() {
// Construct the message.
let message =
Message::BlockRequest(BlockRequest { start_height: height, end_height: height + 1 });
// Send the message to the peers.
for sync_ip in sync_ips {
// If the send fails for any peer, remove the block request from the sync pool.
if validator.send(sync_ip, message.clone()).is_none() {
// Remove the entire block request.
validator.router.sync().remove_block_request(height);
// Break out of the loop.
break 'outer;
}
}
// Sleep for 10 milliseconds to avoid triggering spam detection.
tokio::time::sleep(Duration::from_millis(10)).await;
}
}
}
}));
Ok(())
}
/// Attempts to advance with blocks from the sync pool.
fn advance_with_sync_blocks(&self) {
// Retrieve the latest block height.
let mut current_height = self.ledger.latest_height();
// Try to advance the ledger with the sync pool.
while let Some(block) = self.router.sync().remove_block_response(current_height + 1) {
// Ensure the block height matches.
if block.height() != current_height + 1 {
warn!("Block height mismatch: expected {}, found {}", current_height + 1, block.height());
break;
}
// Check the next block.
if let Err(error) = self.consensus.check_next_block(&block) {
warn!("The next block ({}) is invalid - {error}", block.height());
break;
}
// Attempt to advance to the next block.
if let Err(error) = self.consensus.advance_to_next_block(&block) {
warn!("{error}");
break;
}
// Insert the height and hash as canon in the sync pool.
self.router.sync().insert_canon_locator(block.height(), block.hash());
// Increment the latest height.
current_height += 1;
}
}
}