[][src]Struct radiate::models::neat::layers::lstm::LSTM

pub struct LSTM {
    pub input_size: u32,
    pub memory_size: u32,
    pub output_size: u32,
    pub activation: Activation,
    pub memory: Vec<f32>,
    pub hidden: Vec<f32>,
    pub states: LSTMState,
    pub g_gate: Arc<RwLock<Dense>>,
    pub i_gate: Arc<RwLock<Dense>>,
    pub f_gate: Arc<RwLock<Dense>>,
    pub o_gate: Arc<RwLock<Dense>>,
    pub v_gate: Arc<RwLock<Dense>>,
}

LSTM is a long-short term memory cell represented by a collection of Dense layers and two distinct memory vectors which get updated and travel 'through time'

Fields

input_size: u32memory_size: u32output_size: u32activation: Activationmemory: Vec<f32>hidden: Vec<f32>states: LSTMStateg_gate: Arc<RwLock<Dense>>i_gate: Arc<RwLock<Dense>>f_gate: Arc<RwLock<Dense>>o_gate: Arc<RwLock<Dense>>v_gate: Arc<RwLock<Dense>>

Methods

impl LSTM[src]

pub fn new(
    input_size: u32,
    memory_size: u32,
    output_size: u32,
    activation: Activation
) -> Self
[src]

pub fn step_forward_async(&mut self, inputs: &Vec<f32>) -> Option<Vec<f32>>[src]

feed forward with each forward propagation being executed in a seperate thread to speed up the forward pass if the network is NOT being evolved, if it is, there are already so many theads working to optimize the entire population that extra threading is obsolute and might actually slow it down

pub fn step_forward(&mut self, inputs: &Vec<f32>) -> Option<Vec<f32>>[src]

step forward syncronously

pub fn step_back(&mut self, errors: &Vec<f32>, l_rate: f32) -> Option<Vec<f32>>[src]

Preform one step backwards for the layer. Set the tracer historical meta data to look at the current index, and use that data to compute the gradient steps for eachweight in each gated network. If update is true, the gates will take the accumulated gradient steps, and add them to their respecive weight values

Trait Implementations

impl Clone for LSTM[src]

Implement clone for the neat neural network in order to facilitate proper crossover and mutation for the network (*self.environment.read().unwrap()).clone();

impl Debug for LSTM[src]

impl<'de> Deserialize<'de> for LSTM[src]

impl Display for LSTM[src]

implement display for the LSTM layer of the network

impl Genome<LSTM, NeatEnvironment> for LSTM where
    LSTM: Layer
[src]

in order for the lstm layer to be evolved along with the rest of the network, Genome must be implemented so that the layer can be crossed over and measured along with other lstm layers

fn crossover(
    child: &LSTM,
    parent_two: &LSTM,
    env: &Arc<RwLock<NeatEnvironment>>,
    crossover_rate: f32
) -> Option<LSTM>
[src]

implement how to crossover two LSTM layers

fn distance(one: &LSTM, two: &LSTM, env: &Arc<RwLock<NeatEnvironment>>) -> f32[src]

get the distance between two LSTM layers of the network

impl Layer for LSTM[src]

fn forward(&mut self, inputs: &Vec<f32>) -> Option<Vec<f32>>[src]

forward propagate inputs, if the model is being evolved don't spawn extra threads because it slows down the process by about double the original time. If the model is being trained traditionally, step forward asynconously by spawnin a thread for each individual gate which results in speeds about double as a synconous thread.

fn backward(
    &mut self,
    errors: &Vec<f32>,
    learning_rate: f32
) -> Option<Vec<f32>>
[src]

apply backpropagation through time asyncronously because this is not done during evolution

fn reset(&mut self)[src]

reset the lstm network by clearing the tracer and the states as well as the memory and hidden state

fn add_tracer(&mut self)[src]

add tracers to all the gate.write().unwrap()s in the layer

fn remove_tracer(&mut self)[src]

remove the tracers from all the gate.write().unwrap()s in the layer

impl Send for LSTM[src]

These must be implemneted for the network or any type to be used within seperate threads. Because implementing the functions themselves is dangerious and unsafe and i'm not smart enough to do that from scratch, these "implmenetaions" will get rid of the error and realistically they don't need to be implemneted for the program to work

impl Serialize for LSTM[src]

impl Sync for LSTM[src]

Auto Trait Implementations

impl RefUnwindSafe for LSTM

impl Unpin for LSTM

impl UnwindSafe for LSTM

Blanket Implementations

impl<T> Any for T where
    T: 'static + ?Sized
[src]

impl<T> Borrow<T> for T where
    T: ?Sized
[src]

impl<T> BorrowMut<T> for T where
    T: ?Sized
[src]

impl<T> DeserializeOwned for T where
    T: Deserialize<'de>, 
[src]

impl<T> From<T> for T[src]

impl<T, U> Into<U> for T where
    U: From<T>, 
[src]

impl<L> LayerClone for L where
    L: 'static + Layer + Clone
[src]

impl<T> Serialize for T where
    T: Serialize + ?Sized
[src]

impl<T> ToOwned for T where
    T: Clone
[src]

type Owned = T

The resulting type after obtaining ownership.

impl<T> ToString for T where
    T: Display + ?Sized
[src]

impl<T, U> TryFrom<U> for T where
    U: Into<T>, 
[src]

type Error = Infallible

The type returned in the event of a conversion error.

impl<T, U> TryInto<U> for T where
    U: TryFrom<T>, 
[src]

type Error = <U as TryFrom<T>>::Error

The type returned in the event of a conversion error.

impl<V, T> VZip<V> for T where
    V: MultiLane<T>,