//! `laddu` (/ˈlʌduː/) is a library for analysis of particle physics data. It is intended to be a simple and efficient alternative to some of the [other tools](#alternatives) out there. `laddu` is written in Rust with bindings to Python via [`PyO3`](https://github.com/PyO3/pyo3) and [`maturin`](https://github.com/PyO3/maturin) and is the spiritual successor to [`rustitude`](https://github.com/denehoffman/rustitude), one of my first Rust projects. The goal of this project is to allow users to perform complex amplitude analyses (like partial-wave analyses) without complex code or configuration files.
//!
//! <div class="warning">
//!
//! This crate is still in an early development phase, and the API is not stable. It can (and likely will) be subject to breaking changes before the 1.0.0 version release (and hopefully not many after that).
//!
//! </div>
//!
//! # Table of Contents
//! - [Key Features](#key-features)
//! - [Installation](#installation)
//! - [Quick Start](#quick-start)
//! - [Writing a New Amplitude](#writing-a-new-amplitude)
//! - [Calculating a Likelihood](#calculating-a-likelihood)
//! - [Data Format](#data-format)
//! - [MPI Support](#mpi-support)
//! - [Future Plans](#future-plans)
//! - [Alternatives](#alternatives)
//!
//! # Key Features
//! * A simple interface focused on combining [`Amplitude`](crate::amplitudes::Amplitude)s into models which can be evaluated over [`Dataset`]s.
//! * A single [`Amplitude`](crate::amplitudes::Amplitude) trait which makes it easy to write new amplitudes and integrate them into the library.
//! * Easy interfaces to precompute and cache values before the main calculation to speed up model evaluations.
//! * Efficient parallelism using [`rayon`](https://github.com/rayon-rs/rayon).
//! * Python bindings to allow users to write quick, easy-to-read code that just works.
//!
//! # Installation
//! `laddu` can be added to a Rust project with `cargo`:
//! ```shell
//! cargo add laddu
//! ```
//!
//! The library's Python bindings are located in a library by the same name, which can be installed simply with your favorite Python package manager:
//! ```shell
//! pip install laddu
//! ```
//! or
//! ```shell
//! uv add laddu
//! ```
//!
//! # Quick Start
//! ## Rust
//! ### Writing a New Amplitude
//! At the time of writing, Rust is not a common language used by particle physics, but this tutorial should hopefully convince the reader that they don't have to know the intricacies of Rust to write performant amplitudes. As an example, here is how one might write a Breit-Wigner, parameterized as follows:
//! ```math
//! I_{\ell}(m; m_0, \Gamma_0, m_1, m_2) = \frac{1}{\pi}\frac{m_0 \Gamma_0 B_{\ell}(m, m_1, m_2)}{(m_0^2 - m^2) - \imath m_0 \Gamma}
//! ```
//! where
//! ```math
//! \Gamma = \Gamma_0 \frac{m_0}{m} \frac{q(m, m_1, m_2)}{q(m_0, m_1, m_2)} \left(\frac{B_{\ell}(m, m_1, m_2)}{B_{\ell}(m_0, m_1, m_2)}\right)^2
//! ```
//! is the relativistic width correction, $`q(m_a, m_b, m_c)`$ is the breakup momentum of a particle with mass $`m_a`$ decaying into two particles with masses $`m_b`$ and $`m_c`$, $`B_{\ell}(m_a, m_b, m_c)`$ is the Blatt-Weisskopf barrier factor for the same decay assuming particle $`a`$ has angular momentum $`\ell`$, $`m_0`$ is the mass of the resonance, $`\Gamma_0`$ is the nominal width of the resonance, $`m_1`$ and $`m_2`$ are the masses of the decay products, and $`m`$ is the "input" mass.
//!
//! Although this particular amplitude is already included in `laddu`, let's assume it isn't and imagine how we would write it from scratch:
//!
//! ```rust,no_run
//! use laddu::{
//! AmplitudeID, Cache, DatasetMetadata, Expression, LadduResult, Mass,
//! ParameterID, Parameter, Parameters, Resources,
//! };
//! use laddu::NamedEventView;
//! use laddu::resources::ScalarID;
//! use laddu::traits::*;
//! use laddu::utils::functions::{BarrierKind, QR_DEFAULT, Sheet, blatt_weisskopf_m, q_m};
//! use laddu::{Deserialize, Serialize, typetag};
//! use num::complex::Complex64;
//!
//! #[derive(Clone, Serialize, Deserialize)]
//! pub struct MyBreitWigner {
//! name: String,
//! mass: Parameter,
//! width: Parameter,
//! pid_mass: ParameterID,
//! pid_width: ParameterID,
//! l: usize,
//! daughter_1_mass: Mass,
//! daughter_2_mass: Mass,
//! resonance_mass: Mass,
//! daughter_1_mass_id: ScalarID,
//! daughter_2_mass_id: ScalarID,
//! resonance_mass_id: ScalarID,
//! }
//! impl MyBreitWigner {
//! pub fn new(
//! name: &str,
//! mass: Parameter,
//! width: Parameter,
//! l: usize,
//! daughter_1_mass: &Mass,
//! daughter_2_mass: &Mass,
//! resonance_mass: &Mass,
//! ) -> LadduResult<Expression> {
//! Self {
//! name: name.to_string(),
//! mass,
//! width,
//! pid_mass: ParameterID::default(),
//! pid_width: ParameterID::default(),
//! l,
//! daughter_1_mass: daughter_1_mass.clone(),
//! daughter_2_mass: daughter_2_mass.clone(),
//! resonance_mass: resonance_mass.clone(),
//! daughter_1_mass_id: ScalarID::default(),
//! daughter_2_mass_id: ScalarID::default(),
//! resonance_mass_id: ScalarID::default(),
//! }
//! .into_expression()
//! }
//! }
//!
//! #[typetag::serde]
//! impl Amplitude for MyBreitWigner {
//! fn register(&mut self, resources: &mut Resources) -> LadduResult<AmplitudeID> {
//! self.pid_mass = resources.register_parameter(&self.mass)?;
//! self.pid_width = resources.register_parameter(&self.width)?;
//! self.daughter_1_mass_id = resources.register_scalar(Some(&format!("{}.daughter_1_mass", self.name)));
//! self.daughter_2_mass_id = resources.register_scalar(Some(&format!("{}.daughter_2_mass", self.name)));
//! self.resonance_mass_id = resources.register_scalar(Some(&format!("{}.resonance_mass", self.name)));
//! resources.register_amplitude(&self.name)
//! }
//!
//! fn bind(
//! &mut self,
//! metadata: &DatasetMetadata,
//! ) -> LadduResult<()> {
//! self.daughter_1_mass.bind(metadata)?;
//! self.daughter_2_mass.bind(metadata)?;
//! self.resonance_mass.bind(metadata)?;
//! Ok(())
//! }
//!
//! fn precompute(&self, event: &NamedEventView<'_>, cache: &mut Cache) {
//! cache.store_scalar(self.daughter_1_mass_id, event.evaluate(&self.daughter_1_mass));
//! cache.store_scalar(self.daughter_2_mass_id, event.evaluate(&self.daughter_2_mass));
//! cache.store_scalar(self.resonance_mass_id, event.evaluate(&self.resonance_mass));
//! }
//!
//! fn compute(&self, parameters: &Parameters, cache: &Cache) -> Complex64 {
//! let m = cache.get_scalar(self.resonance_mass_id);
//! let m0 = parameters.get(self.pid_mass).abs();
//! let width0 = parameters.get(self.pid_width).abs();
//! let m1 = cache.get_scalar(self.daughter_1_mass_id);
//! let m2 = cache.get_scalar(self.daughter_2_mass_id);
//! let q0 = q_m(m0, m1, m2, Sheet::Physical);
//! let q = q_m(m, m1, m2, Sheet::Physical);
//! let b0 = blatt_weisskopf_m(m0, m1, m2, self.l, QR_DEFAULT, Sheet::Physical, BarrierKind::Full);
//! let b = blatt_weisskopf_m(m, m1, m2, self.l, QR_DEFAULT, Sheet::Physical, BarrierKind::Full);
//! let width = width0 * (m0 / m) * (q / q0) * (b / b0).powi(2);
//! 1.0 / (Complex64::from(m0.powi(2) - m.powi(2)) - Complex64::I * m0 * width)
//! }
//! }
//! ```
//!
//! While it isn't shown here, we can often be more efficient when implementing
//! [`Amplitude`](crate::amplitudes::Amplitude)s by precomputing values which do not depend on the
//! free parameters. See the [`Amplitude::precompute`](crate::amplitudes::Amplitude::precompute)
//! method for more details.
//!
//! ### Calculating a Likelihood
//! We could then write some code to use this amplitude. For demonstration purposes, let's just calculate an extended unbinned negative log-likelihood, assuming we have some data and Monte Carlo in the proper [parquet format](#data-format):
//! ```rust,no_run
//! use laddu::{Scalar, Dataset, DatasetReadOptions, Mass, NLL, parameter};
//! # use laddu::{
//! # AmplitudeID, Cache, DatasetMetadata, Expression, LadduResult,
//! # ParameterID, Parameter, Parameters, Resources,
//! # };
//! # use laddu::NamedEventView;
//! # use laddu::resources::ScalarID;
//! # use laddu::traits::*;
//! # use laddu::utils::functions::{BarrierKind, QR_DEFAULT, Sheet, blatt_weisskopf_m, q_m};
//! # use laddu::{Deserialize, Serialize, typetag};
//! # use num::complex::Complex64;
//! #
//! # #[derive(Clone, Serialize, Deserialize)]
//! # pub struct MyBreitWigner {
//! # name: String,
//! # mass: Parameter,
//! # width: Parameter,
//! # pid_mass: ParameterID,
//! # pid_width: ParameterID,
//! # l: usize,
//! # daughter_1_mass: Mass,
//! # daughter_2_mass: Mass,
//! # resonance_mass: Mass,
//! # daughter_1_mass_id: ScalarID,
//! # daughter_2_mass_id: ScalarID,
//! # resonance_mass_id: ScalarID,
//! # }
//! # impl MyBreitWigner {
//! # pub fn new(
//! # name: &str,
//! # mass: Parameter,
//! # width: Parameter,
//! # l: usize,
//! # daughter_1_mass: &Mass,
//! # daughter_2_mass: &Mass,
//! # resonance_mass: &Mass,
//! # ) -> LadduResult<Expression> {
//! # Self {
//! # name: name.to_string(),
//! # mass,
//! # width,
//! # pid_mass: ParameterID::default(),
//! # pid_width: ParameterID::default(),
//! # l,
//! # daughter_1_mass: daughter_1_mass.clone(),
//! # daughter_2_mass: daughter_2_mass.clone(),
//! # resonance_mass: resonance_mass.clone(),
//! # daughter_1_mass_id: ScalarID::default(),
//! # daughter_2_mass_id: ScalarID::default(),
//! # resonance_mass_id: ScalarID::default(),
//! # }
//! # .into_expression()
//! # }
//! # }
//! #
//! # #[typetag::serde]
//! # impl Amplitude for MyBreitWigner {
//! # fn register(&mut self, resources: &mut Resources) -> LadduResult<AmplitudeID> {
//! # self.pid_mass = resources.register_parameter(&self.mass)?;
//! # self.pid_width = resources.register_parameter(&self.width)?;
//! # self.daughter_1_mass_id = resources.register_scalar(Some(&format!("{}.daughter_1_mass", self.name)));
//! # self.daughter_2_mass_id = resources.register_scalar(Some(&format!("{}.daughter_2_mass", self.name)));
//! # self.resonance_mass_id = resources.register_scalar(Some(&format!("{}.resonance_mass", self.name)));
//! # resources.register_amplitude(&self.name)
//! # }
//! #
//! # fn bind(
//! # &mut self,
//! # metadata: &DatasetMetadata,
//! # ) -> LadduResult<()> {
//! # self.daughter_1_mass.bind(metadata)?;
//! # self.daughter_2_mass.bind(metadata)?;
//! # self.resonance_mass.bind(metadata)?;
//! # Ok(())
//! # }
//! #
//! # fn precompute(&self, event: &NamedEventView<'_>, cache: &mut Cache) {
//! # cache.store_scalar(self.daughter_1_mass_id, event.evaluate(&self.daughter_1_mass));
//! # cache.store_scalar(self.daughter_2_mass_id, event.evaluate(&self.daughter_2_mass));
//! # cache.store_scalar(self.resonance_mass_id, event.evaluate(&self.resonance_mass));
//! # }
//! #
//! # fn compute(&self, parameters: &Parameters, cache: &Cache) -> Complex64 {
//! # let m = cache.get_scalar(self.resonance_mass_id);
//! # let m0 = parameters.get(self.pid_mass).abs();
//! # let width0 = parameters.get(self.pid_width).abs();
//! # let m1 = cache.get_scalar(self.daughter_1_mass_id);
//! # let m2 = cache.get_scalar(self.daughter_2_mass_id);
//! # let q0 = q_m(m0, m1, m2, Sheet::Physical);
//! # let q = q_m(m, m1, m2, Sheet::Physical);
//! # let b0 = blatt_weisskopf_m(m0, m1, m2, self.l, QR_DEFAULT, Sheet::Physical, BarrierKind::Full);
//! # let b = blatt_weisskopf_m(m, m1, m2, self.l, QR_DEFAULT, Sheet::Physical, BarrierKind::Full);
//! # let width = width0 * (m0 / m) * (q / q0) * (b / b0).powi(2);
//! # 1.0 / (Complex64::from(m0.powi(2) - m.powi(2)) - Complex64::I * m0 * width)
//! # }
//! # }
//! let p4_names = ["beam", "proton", "kshort1", "kshort2"];
//! let aux_names = ["pol_magnitude", "pol_angle"];
//! let options = DatasetReadOptions::default().p4_names(p4_names).aux_names(aux_names);
//! let ds_data = laddu::io::read_parquet("test_data/data.parquet", &options).unwrap();
//! let ds_mc = laddu::io::read_parquet("test_data/mc.parquet", &options).unwrap();
//!
//! let resonance_mass = Mass::new(["kshort1", "kshort2"]);
//! let p1_mass = Mass::new(["kshort1"]);
//! let p2_mass = Mass::new(["kshort2"]);
//! let bw = MyBreitWigner::new(
//! "bw",
//! parameter!("mass"),
//! parameter!("width"),
//! 2,
//! &p1_mass,
//! &p2_mass,
//! &resonance_mass,
//! ).unwrap();
//! let mag = Scalar::new("mag", parameter!("magnitude")).unwrap();
//! let expr = (mag * bw).norm_sqr();
//!
//! let nll = NLL::new(&expr, &ds_data, &ds_mc, None).unwrap();
//! println!("Parameters names and order: {:?}", nll.parameters());
//! let result = nll.evaluate(&[1.27, 0.120, 100.0]).unwrap();
//! println!("The extended negative log-likelihood is {}", result);
//! ```
//! In practice, amplitudes can also be added together, their real and imaginary parts can be taken, and evaluators should mostly take the real part of whatever complex value comes out of the model.
//!
//! # Data Format
//! The data format for `laddu` is a bit different from some of the alternatives like [`AmpTools`](https://github.com/mashephe/AmpTools). Since ROOT doesn't yet have bindings to Rust and projects to read ROOT files are still largely works in progress (although I hope to use [`oxyroot`](https://github.com/m-dupont/oxyroot) in the future when I can figure out a few bugs), the primary interface for data in `laddu` is Parquet files. These are easily accessible from almost any other language and they don't take up much more space than ROOT files. In the interest of future compatibility with any number of experimental setups, the data format consists of an arbitrary number of columns containing the four-momenta of each particle, optional auxiliary scalars (for example, polarization magnitudes/angles), and a single column for the weight. Four-momenta are described by choosing a unique particle identifier and appending the suffixes `_px`, `_py`, `_pz`, and `_e`. Auxiliary values are listed explicitly by name. All numeric columns may be stored as `Float32` or `Float64`; they are promoted to `f64` on read. For example, the following columns describe a dataset with four particles, the first of which is a polarized photon beam, as in the GlueX experiment:
//! | Column name | Data Type | Interpretation |
//! | ------------- | ------------------ | ----------------------------------- |
//! | `beam_px` | `Float32` or `Float64` | Beam momentum (x-component) |
//! | `beam_py` | `Float32` or `Float64` | Beam momentum (y-component) |
//! | `beam_pz` | `Float32` or `Float64` | Beam momentum (z-component) |
//! | `beam_e` | `Float32` or `Float64` | Beam energy |
//! | `pol_magnitude` | `Float32` or `Float64` | Beam polarization magnitude |
//! | `pol_angle` | `Float32` or `Float64` | Beam polarization angle |
//! | `proton_px` | `Float32` or `Float64` | Recoil proton momentum (x-component) |
//! | `proton_py` | `Float32` or `Float64` | Recoil proton momentum (y-component) |
//! | `proton_pz` | `Float32` or `Float64` | Recoil proton momentum (z-component) |
//! | `proton_e` | `Float32` or `Float64` | Recoil proton energy |
//! | `kaon1_px` | `Float32` or `Float64` | Decay product 1 momentum (x) |
//! | `kaon1_py` | `Float32` or `Float64` | Decay product 1 momentum (y) |
//! | `kaon1_pz` | `Float32` or `Float64` | Decay product 1 momentum (z) |
//! | `kaon1_e` | `Float32` or `Float64` | Decay product 1 energy |
//! | `kaon2_px` | `Float32` or `Float64` | Decay product 2 momentum (x) |
//! | `kaon2_py` | `Float32` or `Float64` | Decay product 2 momentum (y) |
//! | `kaon2_pz` | `Float32` or `Float64` | Decay product 2 momentum (z) |
//! | `kaon2_e` | `Float32` or `Float64` | Decay product 2 energy |
//! | `weight` | `Float32` or `Float64` | Event weight |
//!
//! AmpTools-format ROOT tuples can also be loaded through the Python bindings of `laddu` by calling
//! `laddu.io.read_amptools(...)`, which performs the conversion automatically. The Rust
//! API currently supports Parquet and standard ROOT TTrees.
//!
//! # MPI Support
//!
//! The latest version of `laddu` supports the Message Passing Interface (MPI) protocol for distributed computing. MPI-compatible versions of the core `laddu` methods have been written behind the `mpi` feature gate. To build `laddu` with MPI compatibility, it can be added with the `mpi` feature via `cargo add laddu --features mpi`. Note that this requires a working MPI installation, and [OpenMPI](https://www.open-mpi.org/) or [MPICH](https://www.mpich.org/) are recommended, as well as [LLVM](https://llvm.org/)/[Clang](https://clang.llvm.org/). The installation of these packages differs by system, but are generally available via each system's package manager.
//!
//! To use MPI in Rust, one must simply surround their main analysis code with a call to `laddu::mpi::use_mpi(true)` and `laddu::mpi::finalize_mpi()`. The first method has a boolean flag which allows for runtime switching of MPI use (for example, disabling MPI with an environment variable).
//!
//! <div class="warning">
//!
//! The current ROOT backend always materializes the entire TTree on rank 0 before partitioning, so distributed runs do not yet see the same memory savings they do with Parquet inputs.
//!
//! </div>
//!
//! # Future Plans
//! * GPU integration (this is incredibly difficult to do right now, but it's something I'm looking into).
//! * As always, more tests and documentation.
//!
//! # Alternatives
//! While this is likely the first Rust project (aside from my previous attempt, [`rustitude`](https://github.com/denehoffman/rustitude)), there are several other amplitude analysis programs out there at time of writing. This library is a rewrite of `rustitude` which was written when I was just learning Rust and didn't have a firm grasp of a lot of the core concepts that are required to make the analysis pipeline memory- and CPU-efficient. In particular, `rustitude` worked well, but ate up a ton of memory and did not handle precalculation as nicely.
//!
//! ### AmpTools
//! The main inspiration for this project is the library most of my collaboration uses, [`AmpTools`](https://github.com/mashephe/AmpTools). `AmpTools` has several advantages over `laddu`: it's probably faster for almost every use case, but this is mainly because it is fully integrated with MPI and GPU support. I'm not actually sure if there's a fair benchmark between the two libraries, but I'd wager `AmpTools` would still win. `AmpTools` is a much older, more developed project, dating back to 2010. However, it does have its disadvantages. First and foremost, the primary interaction with the library is through configuration files which are not really code and sort of represent a domain specific language. As such, there isn't really a way to check if a particular config will work before running it. Users could technically code up their analyses in C++ as well, but I think this would generally be more work for very little benefit. AmpTools primarily interacts with Minuit, so there aren't simple ways to perform alternative optimization algorithms, and the outputs are a file which must also be parsed by code written by the user. This usually means some boilerplate setup for each analysis, a slew of input and output files, and, since it doesn't ship with any amplitudes, integration with other libraries. The data format is also very rigid, to the point where including beam polarization information feels hacked on (see the Zlm implementation [here](https://github.com/JeffersonLab/halld_sim/blob/6815c979cac4b79a47e5183cf285ce9589fe4c7f/src/libraries/AMPTOOLS_AMPS/Zlm.cc#L26) which requires the event-by-event polarization to be stored in the beam's four-momentum). While there isn't an official Python interface, Lawrence Ng has made some progress porting the code [here](https://github.com/lan13005/PyAmpTools).
//!
//! ### PyPWA
//! [`PyPWA`](https://github.com/JeffersonLab/PyPWA/tree/main) is a library written in pure Python. While this might seem like an issue for performance (and it sort of is), the library has several features which encourage the use of JIT compilers. The upside is that analyses can be quickly prototyped and run with very few dependencies, it can even run on GPUs and use multiprocessing. The downside is that recent development has been slow and the actual implementation of common amplitudes is, in my opinion, [messy](https://pypwa.jlab.org/AmplitudeTWOsim.py). I don't think that's a reason to not use it, but it does make it difficult for new users to get started.
//!
//! ### ComPWA
//! [`ComPWA`](https://compwa.github.io/) is a newcomer to the field. It's also a pure Python implementation and is comprised of three separate libraries. [`QRules`](https://github.com/ComPWA/qrules) can be used to validate and generate particle reaction topologies using conservation rules. [`AmpForm`](https://github.com/ComPWA/ampform) uses `SymPy` to transform these topologies into mathematical expressions, and it can also simplify the mathematical forms through the built-in CAS of `SymPy`. Finally, [`TensorWaves`](https://github.com/ComPWA/tensorwaves) connects `AmpForm` to various fitting methods. In general, these libraries have tons of neat features, are well-documented, and are really quite nice to use. I would like to eventually see `laddu` as a companion to `ComPWA` (rather than direct competition), but I don't really know enough about the libraries to say much more than that.
//!
//! ### Others
//! It could be the case that I am leaving out software with which I am not familiar. If so, I'd love to include it here for reference. I don't think that `laddu` will ever be the end-all-be-all of amplitude analysis, just an alternative that might improve on existing systems. It is important for physicists to be aware of these alternatives. For example, if you really don't want to learn Rust but need to implement an amplitude which isn't already included here, `laddu` isn't for you, and one of these alternatives might be best.
#![warn(clippy::perf, clippy::style, missing_docs)]
/// Methods for loading and manipulating [`EventData`]-based data.
pub mod data {
pub use laddu_core::data::{
BinnedDataset, Dataset, DatasetMetadata, DatasetReadOptions, DatasetWriteOptions, EventData,
};
}
/// Format-specific IO helpers for [`Dataset`]s.
pub mod io {
pub use laddu_core::data::{read_parquet, read_root, write_parquet, write_root};
}
/// Module for likelihood-related structures and methods
pub mod extensions {
pub use laddu_extensions::*;
}
/// Structures for manipulating the cache and free parameters.
pub mod resources {
pub use laddu_core::resources::*;
}
/// Utility functions, enums, and traits
pub mod utils {
pub use laddu_core::utils::*;
}
/// Useful traits for all crate structs
pub mod traits {
pub use laddu_core::amplitudes::Amplitude;
pub use laddu_core::utils::variables::Variable;
pub use laddu_core::ReadWrite;
pub use laddu_extensions::likelihoods::LikelihoodTerm;
}
/// [`Amplitude`](crate::amplitudes::Amplitude)s and methods for making and evaluating them.
pub mod amplitudes {
pub use laddu_amplitudes::*;
pub use laddu_core::amplitudes::{
Amplitude, AmplitudeID, AmplitudeSemanticField, AmplitudeSemanticKey, Evaluator,
Expression, Parameter,
};
pub use laddu_core::parameter;
}
/// <div class="warning">
///
/// This module contains experimental code which may be untested or unreliable. Use at your own
/// risk! The features contained here may eventually be moved into the standard crate modules.
///
/// </div>
pub mod experimental {
pub use laddu_extensions::experimental::*;
}
pub use laddu_amplitudes::*;
pub use laddu_core::amplitudes::{
AmplitudeID, AmplitudeSemanticField, AmplitudeSemanticKey, Evaluator, Expression, Parameter,
};
pub use laddu_core::data::{
BinnedDataset, Dataset, DatasetMetadata, DatasetReadOptions, DatasetWriteOptions, Event,
EventData, NamedEventView,
};
pub use laddu_core::parameter;
pub use laddu_core::resources::{Cache, ParameterID, Parameters, Resources};
pub use laddu_core::utils::reaction::{
Decay, Particle, ParticleSource, Reaction, ReactionTopology, ResolvedTwoToTwo, TwoToTwoReaction,
};
pub use laddu_core::utils::variables::{
AngleVariables, Angles, CosTheta, Mandelstam, Mass, Phi, PolAngle, PolMagnitude, Polarization,
};
pub use laddu_core::utils::vectors::{Vec3, Vec4};
pub use laddu_core::PI;
pub use laddu_core::{LadduError, LadduResult};
pub use laddu_extensions::*;
pub use serde::{Deserialize, Serialize};
pub use typetag;
/// MPI backend for `laddu`
///
/// Message Passing Interface (MPI) is a protocol which enables communication between multiple
/// CPUs in a high-performance computing environment. While [`rayon`] can parallelize tasks on a
/// single CPU, MPI can also parallelize tasks on multiple CPUs by running independent
/// processes on all CPUs at once (tasks) which are assigned ids (ranks) which tell each
/// process what to do and where to send results. This backend coordinates processes which would
/// typically be parallelized over the events in a [`Dataset`].
///
/// To use this backend, the library must be built with the `mpi` feature, which requires an
/// existing implementation of MPI like OpenMPI or MPICH. All processing code should be
/// sandwiched between calls to [`use_mpi`] and [`finalize_mpi`]:
/// ```ignore
/// fn main() {
/// laddu_core::mpi::use_mpi(true);
/// // laddu analysis code here
/// laddu_core::mpi::finalize_mpi();
/// }
/// ```
///
/// [`finalize_mpi`] must be called to trigger all the methods which clean up the MPI
/// environment. While these are called by default when the [`Universe`](`mpi::environment::Universe`) is dropped, `laddu` uses a static `Universe` that can be accessed by all of the methods that need it, rather than passing the context to each method. This simplifies the way programs can be converted to use MPI, but means that the `Universe` is not automatically dropped at the end of the program (so it must be dropped manually).
#[cfg(feature = "mpi")]
pub mod mpi {
pub use laddu_core::mpi::{
finalize_mpi, get_rank, get_size, get_world, is_root, use_mpi, using_mpi, ROOT_RANK,
};
}