Expand description
A high-performance, discrete-event computation framework for system simulation.
NeXosim (né Asynchronix) is a developer-friendly, yet highly optimized software simulator able to scale to very large simulation with complex time-driven state machines.
It promotes a component-oriented architecture that is familiar to system engineers and closely resembles flow-based programming: a model is essentially an isolated entity with a fixed set of typed inputs and outputs, communicating with other models through message passing via connections defined during bench assembly. Unlike in conventional flow-based programming, request-reply patterns are also possible.
NeXosim leverages asynchronous programming to perform auto-parallelization in a manner that is fully transparent to model authors and users, achieving high computational throughput on large simulation benches by means of a custom multi-threaded executor.
§A practical overview
Simulating a system typically involves three distinct activities:
- the design of simulation models for each sub-system,
- the assembly of a simulation bench from a set of models, performed by inter-connecting model ports,
- the execution of the simulation, managed through periodical increments of the simulation time and by exchange of messages with simulation models.
The following sections go through each of these activities in more details.
§Authoring models
Models can contain four kinds of ports:
- output ports, which are instances of the
Outputtype and can be used to broadcast a message, - requestor ports, which are instances of the
RequestororUniRequestortypes and can be used to broadcast a message and receive an iterator yielding the replies from all connected replier ports, - input ports, which are synchronous or asynchronous methods that
implement the
InputFntrait and take an&mut selfargument, a message argument, and an optional&mut Contextargument, - replier ports, which are similar to input ports but implement the
ReplierFntrait and return a reply.
Messages that are broadcast by an output port to an input port are referred to as events, while messages exchanged between requestor and replier ports are referred to as requests and replies.
Models must implement the Model trait. The main purpose of
this trait is to allow models to specify a
Model::init method that is guaranteed to run once
and only once when the simulation is initialized, i.e. after all models
have been connected but before the simulation starts.
The Model::init methods has a default
implementations, so models that do not require setup and initialization can
simply implement the trait with a one-liner such as impl Model for MyModel {}.
More complex models can be built with the ProtoModel
trait. The ProtoModel::build method makes it
possible to:
- build the final
Modelfrom a builder (the model prototype), - perform possibly blocking actions when the model is added to the simulation rather than when the simulation starts, such as establishing a network connection or configuring hardware devices,
- connect submodels and add them to the simulation.
§A simple model
Let us consider for illustration a simple model that forwards its input after multiplying it by 2. This model has only one input and one output port:
┌────────────┐
│ │
Input ●───────►│ Multiplier ├───────► Output
f64 │ │ f64
└────────────┘Multiplier could be implemented as follows:
use nexosim::model::Model;
use nexosim::ports::Output;
#[derive(Default)]
pub struct Multiplier {
pub output: Output<f64>,
}
impl Multiplier {
pub async fn input(&mut self, value: f64) {
self.output.send(2.0 * value).await;
}
}
impl Model for Multiplier {}§A model using the local context
Models frequently need to schedule actions at a future time or simply get access to the current simulation time. To do so, input and replier methods can take an optional argument that gives them access to a local context.
To show how the local context can be used in practice, let us implement
Delay, a model which simply forwards its input unmodified after a 1s
delay:
use std::time::Duration;
use nexosim::model::{Context, Model};
use nexosim::ports::Output;
#[derive(Default)]
pub struct Delay {
pub output: Output<f64>,
}
impl Delay {
pub fn input(&mut self, value: f64, cx: &mut Context<Self>) {
cx.schedule_event(Duration::from_secs(1), Self::send, value).unwrap();
}
async fn send(&mut self, value: f64) {
self.output.send(value).await;
}
}
impl Model for Delay {}§Assembling simulation benches
A simulation bench is a system of inter-connected models that have been migrated to a simulation.
The assembly process usually starts with the instantiation of models and the
creation of a Mailbox for each model. A mailbox is
essentially a fixed-capacity buffer for events and requests. While each
model has only one mailbox, it is possible to create an arbitrary number of
Addresses pointing to that mailbox.
Addresses are used among others to connect models: each output or requestor
port has a connect method that takes as argument a function pointer to
the corresponding input or replier port method and the address of the
targeted model.
Once all models are connected, they are added to a
SimInit instance, which is a builder type for the
final Simulation.
The easiest way to understand the assembly step is with a short example. Say that we want to assemble the following system from the models implemented above:
┌────────────┐
│ │
┌──►│ Delay ├──┐
┌────────────┐ │ │ │ │ ┌────────────┐
│ │ │ └────────────┘ │ │ │
Input ●──►│ Multiplier ├───┤ ├──►│ Delay ├──► Output
│ │ │ ┌────────────┐ │ │ │
└────────────┘ │ │ │ │ └────────────┘
└──►│ Multiplier ├──┘
│ │
└────────────┘Here is how this could be done:
use std::time::Duration;
use nexosim::ports::EventSlot;
use nexosim::simulation::{Mailbox, SimInit};
use nexosim::time::MonotonicTime;
use models::{Delay, Multiplier};
// Instantiate models.
let mut multiplier1 = Multiplier::default();
let mut multiplier2 = Multiplier::default();
let mut delay1 = Delay::default();
let mut delay2 = Delay::default();
// Instantiate mailboxes.
let multiplier1_mbox = Mailbox::new();
let multiplier2_mbox = Mailbox::new();
let delay1_mbox = Mailbox::new();
let delay2_mbox = Mailbox::new();
// Connect the models.
multiplier1.output.connect(Delay::input, &delay1_mbox);
multiplier1.output.connect(Multiplier::input, &multiplier2_mbox);
multiplier2.output.connect(Delay::input, &delay2_mbox);
delay1.output.connect(Delay::input, &delay2_mbox);
// Keep handles to the system input and output for the simulation.
let mut output_slot = EventSlot::new();
delay2.output.connect_sink(&output_slot);
let input_address = multiplier1_mbox.address();
// Pick an arbitrary simulation start time and build the simulation.
let t0 = MonotonicTime::EPOCH;
let (mut simu, scheduler) = SimInit::new()
.add_model(multiplier1, multiplier1_mbox, "multiplier1")
.add_model(multiplier2, multiplier2_mbox, "multiplier2")
.add_model(delay1, delay1_mbox, "delay1")
.add_model(delay2, delay2_mbox, "delay2")
.init(t0)?;
§Running simulations
The simulation can be controlled in several ways:
- by advancing time, either until the next scheduled event with
Simulation::step, until a specific deadline withSimulation::step_until, or until there are no more scheduled events withSimulation::step_unbounded. - by sending events or queries without advancing simulation time, using
Simulation::process_eventorSimulation::send_query, - by scheduling events with a
Scheduler.
When initialized with the default clock, the simulation will run as fast as
possible, without regard for the actual wall clock time. Alternatively, the
simulation time can be synchronized to the wall clock time using
SimInit::set_clock and providing a
custom Clock type or a readily-available real-time clock
such as AutoSystemClock.
Simulation outputs can be monitored using EventSlots,
EventQueues, or any implementer of the
EventSink trait, connected to one or several model
output ports.
This is an example of simulation that could be performed using the above bench assembly:
// Send a value to the first multiplier.
simu.process_event(Multiplier::input, 21.0, &input_address)?;
// The simulation is still at t0 so nothing is expected at the output of the
// second delay gate.
assert!(output_slot.next().is_none());
// Advance simulation time until the next event and check the time and output.
simu.step()?;
assert_eq!(simu.time(), t0 + Duration::from_secs(1));
assert_eq!(output_slot.next(), Some(84.0));
// Get the answer to the ultimate question of life, the universe & everything.
simu.step()?;
assert_eq!(simu.time(), t0 + Duration::from_secs(2));
assert_eq!(output_slot.next(), Some(42.0));
§Message ordering guarantees
The NeXosim runtime is based on the actor model, meaning that every simulation model can be thought of as an isolated entity running in its own thread. While in practice the runtime will actually multiplex and migrate models over a fixed set of kernel threads, models will indeed run in parallel whenever possible.
Since NeXosim is a time-based simulator, the runtime will always execute tasks in chronological order, thus eliminating most ordering ambiguities that could result from parallel execution. Nevertheless, it is sometimes possible for events and queries generated in the same time slice to lead to ambiguous execution orders. In order to make it easier to reason about such situations, NeXosim provides a set of guarantees about message delivery order. Borrowing from the Pony programming language, we refer to this contract as causal messaging, a property that can be summarized by these two rules:
- one-to-one message ordering guarantee: if model
Asends two events or queriesM1and thenM2to modelB, thenBwill always processM1beforeM2, - transitivity guarantee: if
AsendsM1toBand thenM2toCwhich in turn sendsM3toB, even thoughM1andM2may be processed in any order byBandC, it is guaranteed thatBwill processM1beforeM3.
Both guarantees also extend to same-time events scheduled from the global
Scheduler, i.e. the relative ordering of events
scheduled for the same time is preserved and warranties 1 and 2 above
accordingly hold (assuming model A stands for the scheduler). Likewise,
the relative order of same-time events self-scheduled by a model using its
Context is preserved.
§Cargo feature flags
§Tracing
The tracing feature flag provides support for the
tracing crate and can be
activated in Cargo.toml with:
[dependencies]
nexosim = { version = "0.3.4", features = ["tracing"] }See the tracing module for more information.
§Server
The server feature provides a gRPC server for remote control and monitoring,
e.g. from a Python client. It can be activated with:
[dependencies]
nexosim = { version = "0.3.4", features = ["server"] }See the registry and server modules for more information.
Front-end usage documentation will be added upon release of the NeXosim Python client.
§Other resources
§Other examples
Several examples are available that contain more fleshed
out examples and demonstrate various capabilities of the simulation
framework.
§Other features and advanced topics
While the above overview does cover most basic concepts, more information is available in the modules’ documentation:
- the
modelmodule provides more details about models, model prototypes and hierarchical models; be sure to check as well the documentation ofmodel::Contextfor topics such as self-scheduling methods and event cancellation, - the
portsmodule discusses in more details model ports and simulation endpoints, as well as the ability to modify and filter messages exchanged between ports; it also providesEventSourceandQuerySourceobjects which can be connected to models just likeOutputandRequestorports, but for use as simulation endpoints. - the
registryandservermodules make it possible to manage and monitor a simulation locally or remotely from a NeXosim Python client, - the
simulationmodule discusses mailbox capacity and pathological situations that may lead to a deadlock, - the
timemodule introduces thetime::MonotonicTimemonotonic timestamp object and simulation clocks. - the
tracingmodule discusses time-stamping and filtering oftracingevents.