use crate::chunking::{chunk_text_semantic, ChunkingConfig, Tokenizer};
use crate::context_layers::ContextLayerGenerator;
use crate::context_uri::ContextUri;
use crate::db::MemoryDatabase;
use crate::embeddings::EmbeddingService;
use crate::types::{
CleanupLogEntry, DirectoryListing, EmbeddingHealth, LayerType, MemoryChunk, MemoryConfig,
MemoryContext, MemoryError, MemoryLayer, MemoryNode, MemoryResult, MemoryRetrievalMeta,
MemorySearchResult, MemoryStats, MemoryTier, NodeType, StoreMessageRequest, TreeNode,
};
use chrono::Utc;
use std::path::Path;
use std::sync::Arc;
use tandem_providers::{MemoryConsolidationConfig, ProviderRegistry};
use tokio::sync::Mutex;
pub struct MemoryManager {
db: Arc<MemoryDatabase>,
embedding_service: Arc<Mutex<EmbeddingService>>,
tokenizer: Tokenizer,
}
impl MemoryManager {
fn is_malformed_database_error(err: &crate::types::MemoryError) -> bool {
err.to_string()
.to_lowercase()
.contains("database disk image is malformed")
}
pub fn db(&self) -> &Arc<MemoryDatabase> {
&self.db
}
pub async fn new(db_path: &Path) -> MemoryResult<Self> {
let db = Arc::new(MemoryDatabase::new(db_path).await?);
let embedding_service = Arc::new(Mutex::new(EmbeddingService::new()));
let tokenizer = Tokenizer::new()?;
Ok(Self {
db,
embedding_service,
tokenizer,
})
}
pub async fn store_message(&self, request: StoreMessageRequest) -> MemoryResult<Vec<String>> {
if self
.db
.ensure_vector_tables_healthy()
.await
.unwrap_or(false)
{
tracing::warn!("Memory vector tables were repaired before storing message chunks");
}
let config = if let Some(ref pid) = request.project_id {
self.db.get_or_create_config(pid).await?
} else {
MemoryConfig::default()
};
let chunking_config = ChunkingConfig {
chunk_size: config.chunk_size as usize,
chunk_overlap: config.chunk_overlap as usize,
separator: None,
};
let text_chunks = chunk_text_semantic(&request.content, &chunking_config)?;
if text_chunks.is_empty() {
return Ok(Vec::new());
}
let mut chunk_ids = Vec::with_capacity(text_chunks.len());
let embedding_service = self.embedding_service.lock().await;
for text_chunk in text_chunks {
let chunk_id = uuid::Uuid::new_v4().to_string();
let embedding = embedding_service.embed(&text_chunk.content).await?;
let chunk = MemoryChunk {
id: chunk_id.clone(),
content: text_chunk.content,
tier: request.tier,
session_id: request.session_id.clone(),
project_id: request.project_id.clone(),
source: request.source.clone(),
source_path: request.source_path.clone(),
source_mtime: request.source_mtime,
source_size: request.source_size,
source_hash: request.source_hash.clone(),
created_at: Utc::now(),
token_count: text_chunk.token_count as i64,
metadata: request.metadata.clone(),
};
if let Err(err) = self.db.store_chunk(&chunk, &embedding).await {
tracing::warn!("Failed to store memory chunk {}: {}", chunk.id, err);
let repaired = self.db.try_repair_after_error(&err).await.unwrap_or(false)
|| self
.db
.ensure_vector_tables_healthy()
.await
.unwrap_or(false);
if repaired {
tracing::warn!(
"Retrying memory chunk insert after vector table repair: {}",
chunk.id
);
if let Err(retry_err) = self.db.store_chunk(&chunk, &embedding).await {
if Self::is_malformed_database_error(&retry_err) {
tracing::warn!(
"Memory DB still malformed after vector repair. Resetting memory tables and retrying chunk insert: {}",
chunk.id
);
self.db.reset_all_memory_tables().await?;
self.db.store_chunk(&chunk, &embedding).await?;
} else {
return Err(retry_err);
}
}
} else {
return Err(err);
}
}
chunk_ids.push(chunk_id);
}
if config.auto_cleanup {
self.maybe_cleanup(&request.project_id).await?;
}
Ok(chunk_ids)
}
pub async fn search(
&self,
query: &str,
tier: Option<MemoryTier>,
project_id: Option<&str>,
session_id: Option<&str>,
limit: Option<i64>,
) -> MemoryResult<Vec<MemorySearchResult>> {
let effective_limit = limit.unwrap_or(5);
let embedding_service = self.embedding_service.lock().await;
let query_embedding = embedding_service.embed(query).await?;
drop(embedding_service);
let mut results = Vec::new();
let tiers_to_search = match tier {
Some(t) => vec![t],
None => {
if project_id.is_some() {
vec![MemoryTier::Session, MemoryTier::Project, MemoryTier::Global]
} else {
vec![MemoryTier::Session, MemoryTier::Global]
}
}
};
for search_tier in tiers_to_search {
let tier_results = match self
.db
.search_similar(
&query_embedding,
search_tier,
project_id,
session_id,
effective_limit,
)
.await
{
Ok(results) => results,
Err(err) => {
tracing::warn!(
"Memory tier search failed for {:?}: {}. Attempting vector repair.",
search_tier,
err
);
let repaired = self.db.try_repair_after_error(&err).await.unwrap_or(false)
|| self
.db
.ensure_vector_tables_healthy()
.await
.unwrap_or(false);
if repaired {
match self
.db
.search_similar(
&query_embedding,
search_tier,
project_id,
session_id,
effective_limit,
)
.await
{
Ok(results) => results,
Err(retry_err) => {
tracing::warn!(
"Memory tier search still failing for {:?} after repair: {}",
search_tier,
retry_err
);
continue;
}
}
} else {
continue;
}
}
};
for (chunk, distance) in tier_results {
let similarity = 1.0 - distance.clamp(0.0, 1.0);
results.push(MemorySearchResult { chunk, similarity });
}
}
results.sort_by(|a, b| b.similarity.partial_cmp(&a.similarity).unwrap());
results.truncate(effective_limit as usize);
Ok(results)
}
pub async fn retrieve_context(
&self,
query: &str,
project_id: Option<&str>,
session_id: Option<&str>,
token_budget: Option<i64>,
) -> MemoryResult<MemoryContext> {
let (context, _) = self
.retrieve_context_with_meta(query, project_id, session_id, token_budget)
.await?;
Ok(context)
}
pub async fn retrieve_context_with_meta(
&self,
query: &str,
project_id: Option<&str>,
session_id: Option<&str>,
token_budget: Option<i64>,
) -> MemoryResult<(MemoryContext, MemoryRetrievalMeta)> {
let config = if let Some(pid) = project_id {
self.db.get_or_create_config(pid).await?
} else {
MemoryConfig::default()
};
let budget = token_budget.unwrap_or(config.token_budget);
let retrieval_limit = config.retrieval_k.max(1);
let current_session = if let Some(sid) = session_id {
self.db.get_session_chunks(sid).await?
} else {
Vec::new()
};
let search_results = self
.search(query, None, project_id, session_id, Some(retrieval_limit))
.await?;
let mut score_min: Option<f64> = None;
let mut score_max: Option<f64> = None;
for result in &search_results {
score_min = Some(match score_min {
Some(current) => current.min(result.similarity),
None => result.similarity,
});
score_max = Some(match score_max {
Some(current) => current.max(result.similarity),
None => result.similarity,
});
}
let mut current_session = current_session;
let mut relevant_history = Vec::new();
let mut project_facts = Vec::new();
for result in search_results {
match result.chunk.tier {
MemoryTier::Project => {
project_facts.push(result.chunk);
}
MemoryTier::Global => {
project_facts.push(result.chunk);
}
MemoryTier::Session => {
if !current_session.iter().any(|c| c.id == result.chunk.id) {
relevant_history.push(result.chunk);
}
}
}
}
let mut total_tokens: i64 = current_session.iter().map(|c| c.token_count).sum();
total_tokens += relevant_history.iter().map(|c| c.token_count).sum::<i64>();
total_tokens += project_facts.iter().map(|c| c.token_count).sum::<i64>();
if total_tokens > budget {
let excess = total_tokens - budget;
self.trim_context(
&mut current_session,
&mut relevant_history,
&mut project_facts,
excess,
)?;
total_tokens = current_session.iter().map(|c| c.token_count).sum::<i64>()
+ relevant_history.iter().map(|c| c.token_count).sum::<i64>()
+ project_facts.iter().map(|c| c.token_count).sum::<i64>();
}
let context = MemoryContext {
current_session,
relevant_history,
project_facts,
total_tokens,
};
let chunks_total = context.current_session.len()
+ context.relevant_history.len()
+ context.project_facts.len();
let meta = MemoryRetrievalMeta {
used: chunks_total > 0,
chunks_total,
session_chunks: context.current_session.len(),
history_chunks: context.relevant_history.len(),
project_fact_chunks: context.project_facts.len(),
score_min,
score_max,
};
Ok((context, meta))
}
fn trim_context(
&self,
current_session: &mut Vec<MemoryChunk>,
relevant_history: &mut Vec<MemoryChunk>,
project_facts: &mut Vec<MemoryChunk>,
excess_tokens: i64,
) -> MemoryResult<()> {
let mut tokens_to_remove = excess_tokens;
while tokens_to_remove > 0 && !relevant_history.is_empty() {
if let Some(chunk) = relevant_history.pop() {
tokens_to_remove -= chunk.token_count;
}
}
while tokens_to_remove > 0 && !project_facts.is_empty() {
if let Some(chunk) = project_facts.pop() {
tokens_to_remove -= chunk.token_count;
}
}
while tokens_to_remove > 0 && !current_session.is_empty() {
if let Some(chunk) = current_session.pop() {
tokens_to_remove -= chunk.token_count;
}
}
Ok(())
}
pub async fn clear_session(&self, session_id: &str) -> MemoryResult<u64> {
let count = self.db.clear_session_memory(session_id).await?;
self.db
.log_cleanup(
"manual",
MemoryTier::Session,
None,
Some(session_id),
count as i64,
0,
)
.await?;
Ok(count)
}
pub async fn clear_project(&self, project_id: &str) -> MemoryResult<u64> {
let count = self.db.clear_project_memory(project_id).await?;
self.db
.log_cleanup(
"manual",
MemoryTier::Project,
Some(project_id),
None,
count as i64,
0,
)
.await?;
Ok(count)
}
pub async fn get_stats(&self) -> MemoryResult<MemoryStats> {
self.db.get_stats().await
}
pub async fn get_config(&self, project_id: &str) -> MemoryResult<MemoryConfig> {
self.db.get_or_create_config(project_id).await
}
pub async fn set_config(&self, project_id: &str, config: &MemoryConfig) -> MemoryResult<()> {
self.db.update_config(project_id, config).await
}
pub async fn resolve_uri(&self, uri: &str) -> MemoryResult<Option<MemoryNode>> {
self.db.get_node_by_uri(uri).await
}
pub async fn list_directory(&self, uri: &str) -> MemoryResult<DirectoryListing> {
let nodes = self.db.list_directory(uri).await?;
let directories: Vec<MemoryNode> = nodes
.iter()
.filter(|n| n.node_type == NodeType::Directory)
.cloned()
.collect();
let files: Vec<MemoryNode> = nodes
.iter()
.filter(|n| n.node_type == NodeType::File)
.cloned()
.collect();
Ok(DirectoryListing {
uri: uri.to_string(),
nodes,
total_children: directories.len() + files.len(),
directories,
files,
})
}
pub async fn tree(&self, uri: &str, max_depth: usize) -> MemoryResult<Vec<TreeNode>> {
self.db.get_children_tree(uri, max_depth).await
}
pub async fn create_context_node(
&self,
uri: &str,
node_type: NodeType,
metadata: Option<serde_json::Value>,
) -> MemoryResult<String> {
let parsed_uri =
ContextUri::parse(uri).map_err(|e| MemoryError::InvalidConfig(e.message))?;
let parent_uri = parsed_uri.parent().map(|p| p.to_string());
self.db
.create_node(uri, parent_uri.as_deref(), node_type, metadata.as_ref())
.await
}
pub async fn get_context_layer(
&self,
node_id: &str,
layer_type: LayerType,
) -> MemoryResult<Option<MemoryLayer>> {
self.db.get_layer(node_id, layer_type).await
}
pub async fn store_content_with_layers(
&self,
uri: &str,
content: &str,
metadata: Option<serde_json::Value>,
) -> MemoryResult<String> {
let parsed_uri =
ContextUri::parse(uri).map_err(|e| MemoryError::InvalidConfig(e.message))?;
let node_type = if parsed_uri
.last_segment()
.map(|s| s.ends_with(".md") || s.ends_with(".txt") || s.contains("."))
.unwrap_or(false)
{
NodeType::File
} else {
NodeType::Directory
};
let parent_uri = parsed_uri.parent().map(|p| p.to_string());
let node_id = self
.db
.create_node(uri, parent_uri.as_deref(), node_type, metadata.as_ref())
.await?;
let token_count = self.tokenizer.count_tokens(content) as i64;
self.db
.create_layer(&node_id, LayerType::L2, content, token_count, None)
.await?;
Ok(node_id)
}
pub async fn generate_layers_for_node(
&self,
node_id: &str,
providers: &ProviderRegistry,
) -> MemoryResult<()> {
let l2_layer = self.db.get_layer(node_id, LayerType::L2).await?;
let l2_content = match l2_layer {
Some(layer) => layer.content,
None => return Ok(()),
};
let generator = ContextLayerGenerator::new(Arc::new(providers.clone()));
let (l0_content, l1_content) = generator.generate_layers(&l2_content).await?;
let l0_tokens = self.tokenizer.count_tokens(&l0_content) as i64;
let l1_tokens = self.tokenizer.count_tokens(&l1_content) as i64;
if self.db.get_layer(node_id, LayerType::L0).await?.is_none() {
self.db
.create_layer(node_id, LayerType::L0, &l0_content, l0_tokens, None)
.await?;
}
if self.db.get_layer(node_id, LayerType::L1).await?.is_none() {
self.db
.create_layer(node_id, LayerType::L1, &l1_content, l1_tokens, None)
.await?;
}
Ok(())
}
pub async fn get_layer_content(
&self,
node_id: &str,
layer_type: LayerType,
) -> MemoryResult<Option<String>> {
let layer = self.db.get_layer(node_id, layer_type).await?;
Ok(layer.map(|l| l.content))
}
pub async fn store_content_with_layers_auto(
&self,
uri: &str,
content: &str,
metadata: Option<serde_json::Value>,
providers: Option<&ProviderRegistry>,
) -> MemoryResult<String> {
let node_id = self
.store_content_with_layers(uri, content, metadata)
.await?;
if let Some(p) = providers {
if let Err(e) = self.generate_layers_for_node(&node_id, p).await {
tracing::warn!("Failed to generate layers for node {}: {}", node_id, e);
}
}
Ok(node_id)
}
pub async fn run_cleanup(&self, project_id: Option<&str>) -> MemoryResult<u64> {
let mut total_cleaned = 0u64;
if let Some(pid) = project_id {
let config = self.db.get_or_create_config(pid).await?;
if config.auto_cleanup {
let cleaned = self
.db
.cleanup_old_sessions(config.session_retention_days)
.await?;
total_cleaned += cleaned;
if cleaned > 0 {
self.db
.log_cleanup(
"auto",
MemoryTier::Session,
Some(pid),
None,
cleaned as i64,
0,
)
.await?;
}
}
} else {
let cleaned = self.db.cleanup_old_sessions(30).await?;
total_cleaned += cleaned;
}
if total_cleaned > 100 {
self.db.vacuum().await?;
}
Ok(total_cleaned)
}
async fn maybe_cleanup(&self, project_id: &Option<String>) -> MemoryResult<()> {
if let Some(pid) = project_id {
let stats = self.db.get_stats().await?;
let config = self.db.get_or_create_config(pid).await?;
if stats.project_chunks > config.max_chunks {
let excess = stats.project_chunks - config.max_chunks;
tracing::info!("Project {} has {} excess chunks", pid, excess);
}
}
Ok(())
}
pub async fn get_cleanup_log(&self, _limit: i64) -> MemoryResult<Vec<CleanupLogEntry>> {
Ok(Vec::new())
}
pub fn count_tokens(&self, text: &str) -> usize {
self.tokenizer.count_tokens(text)
}
pub async fn embedding_health(&self) -> EmbeddingHealth {
let service = self.embedding_service.lock().await;
if service.is_available() {
EmbeddingHealth {
status: "ok".to_string(),
reason: None,
}
} else {
EmbeddingHealth {
status: "degraded_disabled".to_string(),
reason: service.disabled_reason().map(ToString::to_string),
}
}
}
pub async fn consolidate_session(
&self,
session_id: &str,
project_id: Option<&str>,
providers: &ProviderRegistry,
config: &MemoryConsolidationConfig,
) -> MemoryResult<Option<String>> {
if !config.enabled {
return Ok(None);
}
let chunks = self.db.get_session_chunks(session_id).await?;
if chunks.is_empty() {
return Ok(None);
}
let mut text_parts = Vec::new();
for chunk in &chunks {
text_parts.push(chunk.content.clone());
}
let full_text = text_parts.join("\n\n---\n\n");
let prompt = format!(
"Please provide a concise but comprehensive summary of the following chat session. \
Focus on the key decisions, technical details, code changes, and unresolved issues. \
Do NOT include conversational filler, greetings, or sign-offs. \
This summary will be used as long-term memory to recall the context of this work.\n\n\
Session transcripts:\n\n{}",
full_text
);
let provider_override = config.provider.as_deref().filter(|s| !s.is_empty());
let model_override = config.model.as_deref().filter(|s| !s.is_empty());
let summary_text = match providers
.complete_cheapest(&prompt, provider_override, model_override)
.await
{
Ok(s) => s,
Err(e) => {
tracing::warn!("Memory consolidation LLM failed for session {session_id}: {e}");
return Ok(None);
}
};
if summary_text.trim().is_empty() {
return Ok(None);
}
let embedding = {
let service = self.embedding_service.lock().await;
service
.embed(&summary_text)
.await
.map_err(|e| crate::types::MemoryError::Embedding(e.to_string()))?
};
let chunk_id = uuid::Uuid::new_v4().to_string();
let chunk = MemoryChunk {
id: chunk_id,
content: summary_text.clone(),
tier: MemoryTier::Project,
session_id: None, project_id: project_id.map(ToString::to_string),
created_at: Utc::now(),
source: "consolidation".to_string(),
token_count: self.count_tokens(&summary_text) as i64,
source_path: None,
source_mtime: None,
source_size: None,
source_hash: None,
metadata: None,
};
self.db.store_chunk(&chunk, &embedding).await?;
self.db.clear_session_memory(session_id).await?;
tracing::info!(
"Session {session_id} consolidated into summary chunk. Original chunks cleared."
);
Ok(Some(summary_text))
}
}
pub async fn create_memory_manager(app_data_dir: &Path) -> MemoryResult<MemoryManager> {
let db_path = app_data_dir.join("tandem_memory.db");
MemoryManager::new(&db_path).await
}
#[cfg(test)]
mod tests {
use super::*;
use tempfile::TempDir;
fn is_embeddings_disabled(err: &crate::types::MemoryError) -> bool {
matches!(err, crate::types::MemoryError::Embedding(msg) if msg.to_ascii_lowercase().contains("embeddings disabled"))
}
async fn setup_test_manager() -> (MemoryManager, TempDir) {
let temp_dir = TempDir::new().unwrap();
let db_path = temp_dir.path().join("test_memory.db");
let manager = MemoryManager::new(&db_path).await.unwrap();
(manager, temp_dir)
}
#[tokio::test]
async fn test_store_and_search() {
let (manager, _temp) = setup_test_manager().await;
let request = StoreMessageRequest {
content: "This is a test message about artificial intelligence and machine learning."
.to_string(),
tier: MemoryTier::Project,
session_id: Some("session-1".to_string()),
project_id: Some("project-1".to_string()),
source: "user_message".to_string(),
source_path: None,
source_mtime: None,
source_size: None,
source_hash: None,
metadata: None,
};
let chunk_ids = match manager.store_message(request).await {
Ok(ids) => ids,
Err(err) if is_embeddings_disabled(&err) => return,
Err(err) => panic!("store_message failed: {err}"),
};
assert!(!chunk_ids.is_empty());
let results = manager
.search(
"artificial intelligence",
None,
Some("project-1"),
None,
None,
)
.await;
let results = match results {
Ok(results) => results,
Err(err) if is_embeddings_disabled(&err) => return,
Err(err) => panic!("search failed: {err}"),
};
assert!(!results.is_empty());
assert!(results[0].similarity >= 0.0);
}
#[tokio::test]
async fn test_retrieve_context() {
let (manager, _temp) = setup_test_manager().await;
let request = StoreMessageRequest {
content: "The project uses React and TypeScript for the frontend.".to_string(),
tier: MemoryTier::Project,
session_id: None,
project_id: Some("project-1".to_string()),
source: "assistant_response".to_string(),
source_path: None,
source_mtime: None,
source_size: None,
source_hash: None,
metadata: None,
};
match manager.store_message(request).await {
Ok(_) => {}
Err(err) if is_embeddings_disabled(&err) => return,
Err(err) => panic!("store_message failed: {err}"),
}
let context = manager
.retrieve_context("What technologies are used?", Some("project-1"), None, None)
.await;
let context = match context {
Ok(context) => context,
Err(err) if is_embeddings_disabled(&err) => return,
Err(err) => panic!("retrieve_context failed: {err}"),
};
assert!(!context.project_facts.is_empty());
}
#[tokio::test]
async fn test_retrieve_context_with_meta() {
let (manager, _temp) = setup_test_manager().await;
let request = StoreMessageRequest {
content: "The backend uses Rust and sqlite-vec for retrieval.".to_string(),
tier: MemoryTier::Project,
session_id: None,
project_id: Some("project-1".to_string()),
source: "assistant_response".to_string(),
source_path: None,
source_mtime: None,
source_size: None,
source_hash: None,
metadata: None,
};
match manager.store_message(request).await {
Ok(_) => {}
Err(err) if is_embeddings_disabled(&err) => return,
Err(err) => panic!("store_message failed: {err}"),
}
let result = manager
.retrieve_context_with_meta("What does the backend use?", Some("project-1"), None, None)
.await;
let (context, meta) = match result {
Ok(v) => v,
Err(err) if is_embeddings_disabled(&err) => return,
Err(err) => panic!("retrieve_context_with_meta failed: {err}"),
};
assert!(meta.chunks_total > 0);
assert!(meta.used);
assert_eq!(
meta.chunks_total,
context.current_session.len()
+ context.relevant_history.len()
+ context.project_facts.len()
);
assert!(meta.score_min.is_some());
assert!(meta.score_max.is_some());
}
#[tokio::test]
async fn test_config_management() {
let (manager, _temp) = setup_test_manager().await;
let config = manager.get_config("project-1").await.unwrap();
assert_eq!(config.max_chunks, 10000);
let new_config = MemoryConfig {
max_chunks: 5000,
retrieval_k: 10,
..Default::default()
};
manager.set_config("project-1", &new_config).await.unwrap();
let updated = manager.get_config("project-1").await.unwrap();
assert_eq!(updated.max_chunks, 5000);
assert_eq!(updated.retrieval_k, 10);
}
}