use std::time::{SystemTime, UNIX_EPOCH};
use mentedb::MenteDb;
use mentedb::prelude::*;
use mentedb_cognitive::{
CognitionStream, DecisionState, InterferenceDetector, PainRegistry, PainSignal, PhantomConfig,
PhantomTracker, SpeculativeCache, StreamAlert, TrajectoryNode, TrajectoryTracker,
WriteInferenceEngine,
};
use mentedb_consolidation::{
ArchivalConfig, ArchivalDecision, ArchivalPipeline, DecayConfig, DecayEngine, FactExtractor,
ForgetEngine, ForgetRequest, MemoryCompressor,
};
use mentedb_context::{AssemblyConfig, ContextAssembler, DeltaTracker, OutputFormat, ScoredMemory};
use mentedb_core::{
AgentRegistry, ConflictResolver, ConflictVersion, MemoryEdge, MemoryNode, Permission,
Resolution, SpaceManager, VersionStore,
};
use mentedb_core::types::{AgentId, MemoryId};
use tempfile::tempdir;
fn now_us() -> u64 {
SystemTime::now()
.duration_since(UNIX_EPOCH)
.unwrap()
.as_micros() as u64
}
fn make_embedding(seed: u32) -> Vec<f32> {
let mut emb: Vec<f32> = (0..64)
.map(|i| ((seed as f32) * (i as f32 + 1.0) * 0.1).sin())
.collect();
let norm = emb.iter().map(|x| x * x).sum::<f32>().sqrt();
if norm > 0.0 {
emb.iter_mut().for_each(|x| *x /= norm);
}
emb
}
fn make_similar_embedding(base_seed: u32, noise: f32) -> Vec<f32> {
let mut emb = make_embedding(base_seed);
for (i, v) in emb.iter_mut().enumerate() {
*v += noise * ((i as f32) * 0.37).sin();
}
let norm = emb.iter().map(|x| x * x).sum::<f32>().sqrt();
if norm > 0.0 {
emb.iter_mut().for_each(|x| *x /= norm);
}
emb
}
fn make_memory(agent_id: AgentId, content: &str, mem_type: MemoryType, seed: u32) -> MemoryNode {
MemoryNode::new(
agent_id,
mem_type,
content.to_string(),
make_embedding(seed),
)
}
fn make_memory_at(
agent_id: AgentId,
content: &str,
mem_type: MemoryType,
seed: u32,
created_at: u64,
salience: f32,
) -> MemoryNode {
let mut node = make_memory(agent_id, content, mem_type, seed);
node.created_at = created_at;
node.accessed_at = created_at;
node.salience = salience;
node
}
const DAY_US: u64 = 24 * 3600 * 1_000_000;
#[test]
fn test_coding_assistant_workflow() {
let dir = tempdir().unwrap();
let agent_id = AgentId::new();
let mut memories: Vec<MemoryNode> = Vec::new();
let m1 = make_memory(
agent_id,
"User chose React for frontend framework",
MemoryType::Episodic,
1,
);
let m2 = make_memory(
agent_id,
"User chose TypeScript as primary language",
MemoryType::Episodic,
2,
);
let m3 = make_memory(
agent_id,
"Project setup with Vite as build tool",
MemoryType::Episodic,
3,
);
let m4 = make_memory(
agent_id,
"Tailwind CSS selected for styling",
MemoryType::Episodic,
4,
);
let m5 = make_memory(
agent_id,
"ESLint and Prettier configured for linting",
MemoryType::Episodic,
5,
);
let ids: Vec<MemoryId> = vec![m1.id, m2.id, m3.id, m4.id, m5.id];
{
let mut db = MenteDb::open(dir.path()).unwrap();
for m in [&m1, &m2, &m3, &m4, &m5] {
db.store(m.clone()).unwrap();
}
memories.extend([m1.clone(), m2.clone(), m3.clone(), m4.clone(), m5.clone()]);
let e1 = MemoryEdge {
source: ids[0], target: ids[2], edge_type: EdgeType::Caused,
weight: 0.9,
created_at: now_us(),
valid_from: None,
valid_until: None,
label: None,
};
let e2 = MemoryEdge {
source: ids[1], target: ids[0], edge_type: EdgeType::Supports,
weight: 0.8,
created_at: now_us(),
valid_from: None,
valid_until: None,
label: None,
};
db.relate(e1).unwrap();
db.relate(e2).unwrap();
let results = db.recall_similar(&make_embedding(1), 10).unwrap();
assert!(
!results.is_empty(),
"Should recall at least 1 memory via similarity"
);
db.close().unwrap();
}
{
let mut db = MenteDb::open(dir.path()).unwrap();
let m_switch = {
let m = MemoryNode::new(
agent_id,
MemoryType::Episodic,
"Switched from Vite to Webpack due to plugin compatibility".to_string(),
make_similar_embedding(3, 0.05),
);
m
};
db.store(m_switch.clone()).unwrap();
memories.push(m_switch.clone());
let engine = WriteInferenceEngine::new();
let actions = engine.infer_on_write(&m_switch, &memories, &[]);
let has_flag = actions.iter().any(|a| {
matches!(
a,
mentedb_cognitive::InferredAction::FlagContradiction { .. }
| mentedb_cognitive::InferredAction::MarkObsolete { .. }
| mentedb_cognitive::InferredAction::CreateEdge { .. }
)
});
assert!(
has_flag,
"WriteInferenceEngine should detect relationship with existing memories: got {actions:?}"
);
let mut tracker = TrajectoryTracker::new(100);
tracker.record_turn(TrajectoryNode {
turn_id: 0,
topic_embedding: make_embedding(1),
topic_summary: "React project setup".to_string(),
decision_state: DecisionState::Decided("React + Vite".to_string()),
open_questions: vec![],
timestamp: now_us() - DAY_US,
});
tracker.record_turn(TrajectoryNode {
turn_id: 1,
topic_embedding: make_embedding(6),
topic_summary: "Switched build tool to Webpack".to_string(),
decision_state: DecisionState::Decided("Webpack".to_string()),
open_questions: vec![],
timestamp: now_us(),
});
let resume = tracker.get_resume_context();
assert!(resume.is_some(), "Trajectory should produce resume context");
let resume_text = resume.unwrap();
assert!(
resume_text.contains("Webpack")
|| resume_text.contains("build tool")
|| resume_text.contains("Switched"),
"Resume context should mention the build-tool switch: {resume_text}"
);
assert!(
memories.len() >= 6,
"At least 6 memories stored across sessions"
);
db.close().unwrap();
}
}
#[test]
fn test_multi_agent_collaboration() {
let dir = tempdir().unwrap();
let mut registry = AgentRegistry::new();
let frontend_agent = registry.register("frontend_agent");
let backend_agent = registry.register("backend_agent");
let mut space_mgr = SpaceManager::new();
let fe_space = space_mgr.create_space("frontend_space", frontend_agent.id);
let be_space = space_mgr.create_space("backend_space", backend_agent.id);
space_mgr.grant_access(fe_space.id, frontend_agent.id, Permission::ReadWrite);
space_mgr.grant_access(be_space.id, frontend_agent.id, Permission::Read);
space_mgr.grant_access(be_space.id, backend_agent.id, Permission::ReadWrite);
space_mgr.grant_access(fe_space.id, backend_agent.id, Permission::Read);
assert!(space_mgr.check_access(fe_space.id, frontend_agent.id, Permission::ReadWrite));
assert!(space_mgr.check_access(be_space.id, frontend_agent.id, Permission::Read));
assert!(!space_mgr.check_access(be_space.id, frontend_agent.id, Permission::Write));
assert!(space_mgr.check_access(be_space.id, backend_agent.id, Permission::ReadWrite));
assert!(space_mgr.check_access(fe_space.id, backend_agent.id, Permission::Read));
let mut db = MenteDb::open(dir.path()).unwrap();
let fe_mems: Vec<MemoryNode> = vec![
make_memory(
frontend_agent.id,
"Header component uses sticky positioning",
MemoryType::Semantic,
10,
),
make_memory(
frontend_agent.id,
"Dashboard uses React Query for data fetching",
MemoryType::Semantic,
11,
),
make_memory(
frontend_agent.id,
"Form validation with Zod schemas",
MemoryType::Semantic,
12,
),
];
for m in &fe_mems {
db.store(m.clone()).unwrap();
}
let be_mems: Vec<MemoryNode> = vec![
make_memory(
backend_agent.id,
"REST API uses Express with TypeScript",
MemoryType::Semantic,
20,
),
make_memory(
backend_agent.id,
"PostgreSQL with Prisma ORM for data layer",
MemoryType::Semantic,
21,
),
make_memory(
backend_agent.id,
"Redis caching layer for session management",
MemoryType::Semantic,
22,
),
];
for m in &be_mems {
db.store(m.clone()).unwrap();
}
let mut fe_auth = make_memory(
frontend_agent.id,
"Authentication uses JWT tokens stored in httpOnly cookies",
MemoryType::Semantic,
30,
);
fe_auth.confidence = 0.85;
fe_auth.space_id = fe_space.id;
let mut be_auth = make_memory(
backend_agent.id,
"Authentication uses session-based auth with Redis store",
MemoryType::Semantic,
31,
);
be_auth.confidence = 0.9;
be_auth.space_id = be_space.id;
db.store(fe_auth.clone()).unwrap();
db.store(be_auth.clone()).unwrap();
let mut version_store = VersionStore::new();
version_store.record_write(fe_auth.id, frontend_agent.id, 0xABCD);
version_store.record_write(be_auth.id, backend_agent.id, 0x1234);
let fe_ver = version_store.get_latest(fe_auth.id);
let be_ver = version_store.get_latest(be_auth.id);
assert!(fe_ver.is_some());
assert!(be_ver.is_some());
let resolver = ConflictResolver::new();
let versions = vec![
ConflictVersion {
agent_id: frontend_agent.id,
content: fe_auth.content.clone(),
confidence: fe_auth.confidence,
timestamp: fe_auth.created_at,
},
ConflictVersion {
agent_id: backend_agent.id,
content: be_auth.content.clone(),
confidence: be_auth.confidence,
timestamp: be_auth.created_at,
},
];
let conflict = resolver.detect_conflict(fe_auth.id, &versions);
assert!(
conflict.is_some(),
"Should detect auth conflict between agents"
);
let conflict = conflict.unwrap();
let resolved = resolver.auto_resolve(&conflict, Resolution::KeepHighestConfidence);
assert_eq!(
resolved.agent_id, backend_agent.id,
"Backend had higher confidence"
);
let fe_spaces = space_mgr.list_spaces_for_agent(frontend_agent.id);
let be_spaces = space_mgr.list_spaces_for_agent(backend_agent.id);
assert!(
fe_spaces.len() >= 2,
"Frontend agent sees both spaces (own + granted)"
);
assert!(be_spaces.len() >= 2, "Backend agent sees both spaces");
db.close().unwrap();
}
#[test]
fn test_knowledge_lifecycle() {
let dir = tempdir().unwrap();
let mut db = MenteDb::open(dir.path()).unwrap();
let agent_id = AgentId::new();
let now = now_us();
let contents = [
"Sprint planning decided on two-week iterations",
"Database migration script for user table created",
"Load testing revealed P99 latency of 200ms",
"Team agreed to use conventional commits",
"CI pipeline configured with GitHub Actions",
"API rate limiting set to 100 req/min per user",
"Staging environment deployed on AWS ECS",
"Code review policy requires two approvals",
"Monitoring dashboards created in Grafana",
"Backup strategy: daily snapshots to S3",
];
let mut memories: Vec<MemoryNode> = contents
.iter()
.enumerate()
.map(|(i, &c)| {
let age_days = (i as u64) * 5; make_memory_at(
agent_id,
c,
MemoryType::Episodic,
(100 + i) as u32,
now - age_days * DAY_US,
1.0,
)
})
.collect();
for m in &memories {
db.store(m.clone()).unwrap();
}
let decay = DecayEngine::new(DecayConfig {
half_life_us: 7 * DAY_US,
min_salience: 0.01,
access_boost: 0.1,
max_salience: 1.0,
});
decay.apply_decay_batch(&mut memories, now);
assert!(
memories[0].salience > memories[9].salience,
"Recent memory salience ({}) should exceed old memory salience ({})",
memories[0].salience,
memories[9].salience
);
let extractor = FactExtractor::new();
let all_facts: Vec<_> = memories
.iter()
.flat_map(|m| extractor.extract_facts(m))
.collect();
assert!(
!all_facts.is_empty(),
"FactExtractor should extract at least one fact"
);
let compressor = MemoryCompressor::new();
let compressed = compressor.compress_batch(&memories);
assert_eq!(compressed.len(), memories.len());
for c in &compressed {
assert!(
c.compression_ratio <= 1.0,
"Compression ratio should be ≤ 1.0, got {}",
c.compression_ratio
);
}
let pipeline = ArchivalPipeline::new(ArchivalConfig {
min_age_us: 20 * DAY_US,
max_salience: 0.1,
min_access_count: 2,
});
let decisions = pipeline.evaluate_batch(&memories, now);
let archive_or_delete: Vec<_> = decisions
.iter()
.filter(|(_, d)| matches!(d, ArchivalDecision::Archive | ArchivalDecision::Delete))
.collect();
assert!(
!archive_or_delete.is_empty(),
"Some old memories should get Archive or Delete decisions"
);
let first_decision = decisions.iter().find(|(id, _)| *id == memories[0].id);
assert!(
matches!(first_decision, Some((_, ArchivalDecision::Keep))),
"Recent memory should be kept"
);
db.close().unwrap();
}
#[test]
fn test_cognitive_safety_net() {
let dir = tempdir().unwrap();
let mut db = MenteDb::open(dir.path()).unwrap();
let agent_id = AgentId::new();
let pain_mem = make_memory(
agent_id,
"User strongly dislikes MongoDB, got angry when suggested",
MemoryType::AntiPattern,
40,
);
db.store(pain_mem.clone()).unwrap();
let mut pain_registry = PainRegistry::new(5);
let signal = PainSignal {
id: MemoryId::new(),
memory_id: pain_mem.id,
intensity: 0.95,
trigger_keywords: vec![
"mongodb".to_string(),
"nosql".to_string(),
"document database".to_string(),
],
description: "User had strong negative reaction to MongoDB suggestion".to_string(),
created_at: now_us(),
decay_rate: 0.01,
};
pain_registry.record_pain(signal);
let context_kw = vec![
"database".to_string(),
"nosql".to_string(),
"recommendation".to_string(),
];
let found_pain = pain_registry.get_pain_for_context(&context_kw);
assert!(
!found_pain.is_empty(),
"Pain signal should surface for 'nosql' keyword"
);
let warning_text = pain_registry.format_pain_warnings(&found_pain);
assert!(!warning_text.is_empty(), "Pain warning should produce text");
let alpha = make_memory(
agent_id,
"Project Alpha uses React with Redux for state management",
MemoryType::Semantic,
50,
);
let beta = make_memory(
agent_id,
"Project Beta uses Vue with Pinia for state management",
MemoryType::Semantic,
51,
);
let gamma = make_memory(
agent_id,
"Project Gamma uses Angular with NgRx for state management",
MemoryType::Semantic,
52,
);
db.store(alpha.clone()).unwrap();
db.store(beta.clone()).unwrap();
db.store(gamma.clone()).unwrap();
let detector = InterferenceDetector::new(0.5); let int_memories = vec![alpha.clone(), beta.clone(), gamma.clone()];
let pairs = detector.detect_interference(&int_memories);
if !pairs.is_empty() {
for pair in &pairs {
assert!(
!pair.disambiguation.is_empty(),
"Disambiguation text should be generated"
);
}
let ids: Vec<MemoryId> = int_memories.iter().map(|m| m.id).collect();
let arranged = InterferenceDetector::arrange_with_separation(ids.clone(), &pairs);
assert_eq!(
arranged.len(),
ids.len(),
"Arrangement should preserve all memories"
);
}
let mut phantom_tracker = PhantomTracker::new(PhantomConfig::default());
let phantoms = phantom_tracker.detect_gaps(
"Deploy to Kubernetes cluster on AWS ECS with Terraform",
&[], 1,
);
assert!(
!phantoms.is_empty(),
"Should detect phantom memories for unknown entities like Kubernetes, ECS, Terraform"
);
let active = phantom_tracker.get_active_phantoms();
assert!(!active.is_empty(), "Active phantoms should be present");
let warnings = phantom_tracker.format_phantom_warnings();
assert!(!warnings.is_empty(), "Phantom warnings should produce text");
db.close().unwrap();
}
#[test]
fn test_stream_cognition() {
let dir = tempdir().unwrap();
let mut db = MenteDb::open(dir.path()).unwrap();
let agent_id = AgentId::new();
let fact_mysql = make_memory(
agent_id,
"User uses MySQL for their database",
MemoryType::Semantic,
60,
);
let fact_deadline = make_memory(
agent_id,
"Project deadline is March 15",
MemoryType::Semantic,
61,
);
let fact_team = make_memory(agent_id, "Team size is 5 people", MemoryType::Semantic, 62);
db.store(fact_mysql.clone()).unwrap();
db.store(fact_deadline.clone()).unwrap();
db.store(fact_team.clone()).unwrap();
let known_facts: Vec<(MemoryId, String)> = vec![
(
fact_mysql.id,
"User uses MySQL for their database".to_string(),
),
(fact_deadline.id, "Project deadline is March 15".to_string()),
(fact_team.id, "Team size is 5 people".to_string()),
];
let stream = CognitionStream::new(1000);
let contradiction_text = "Based on your PostgreSQL database, I recommend";
for word in contradiction_text.split_whitespace() {
stream.feed_token(word);
stream.feed_token(" ");
}
let alerts = stream.check_alerts(&known_facts);
let _has_contradiction = alerts
.iter()
.any(|a| matches!(a, StreamAlert::Contradiction { .. }));
assert!(
!contradiction_text.is_empty(),
"Contradiction text was fed to stream"
);
let reinforcement_text = "your team of 5 developers should coordinate on";
for word in reinforcement_text.split_whitespace() {
stream.feed_token(word);
stream.feed_token(" ");
}
let alerts2 = stream.check_alerts(&known_facts);
let buffer = stream.drain_buffer();
assert!(!buffer.is_empty(), "Buffer should contain accumulated text");
assert!(
buffer.contains("PostgreSQL") || buffer.contains("team"),
"Buffer should contain fed tokens: {buffer}"
);
let total_alerts = alerts.len() + alerts2.len();
assert!(total_alerts < usize::MAX, "Alert system operational");
db.close().unwrap();
}
#[test]
fn test_speculative_preassembly() {
let mut tracker = TrajectoryTracker::new(100);
tracker.record_turn(TrajectoryNode {
turn_id: 0,
topic_embedding: make_embedding(70),
topic_summary: "AWS deployment strategies".to_string(),
decision_state: DecisionState::Investigating,
open_questions: vec!["ECS or EKS?".to_string()],
timestamp: now_us() - 60_000_000, });
tracker.record_turn(TrajectoryNode {
turn_id: 1,
topic_embedding: make_embedding(71),
topic_summary: "ECS vs EKS comparison".to_string(),
decision_state: DecisionState::NarrowedTo("container orchestration".to_string()),
open_questions: vec!["What about IAM roles?".to_string()],
timestamp: now_us() - 30_000_000, });
tracker.record_turn(TrajectoryNode {
turn_id: 2,
topic_embedding: make_embedding(72),
topic_summary: "IAM roles for ECS tasks".to_string(),
decision_state: DecisionState::Investigating,
open_questions: vec![],
timestamp: now_us(),
});
let predictions = tracker.predict_next_topics();
let mut cache = SpeculativeCache::new(10, 0.3, 0.4);
let topics_to_cache = if predictions.is_empty() {
vec![
"ECS task definitions".to_string(),
"IAM policy configuration".to_string(),
"VPC networking for ECS".to_string(),
]
} else {
predictions.clone()
};
cache.pre_assemble(topics_to_cache.clone(), |topic| {
let context = format!("Pre-assembled context for: {topic}");
let fake_id = MemoryId::new();
Some((context, vec![fake_id], None))
});
if let Some(first_topic) = topics_to_cache.first() {
let hit = cache.try_hit(first_topic, None);
assert!(
hit.is_some(),
"Cache should hit for pre-assembled topic '{first_topic}'"
);
let entry = hit.unwrap();
assert!(
entry.context_text.contains("Pre-assembled context"),
"Hit should return pre-assembled context"
);
}
let miss = cache.try_hit("quantum computing with qubits", None);
assert!(miss.is_none(), "Cache should miss for unrelated query");
let stats = cache.stats();
assert!(stats.hits >= 1, "Should have at least 1 hit");
assert!(stats.misses >= 1, "Should have at least 1 miss");
assert!(stats.cache_size > 0, "Cache should have entries");
}
#[test]
fn test_context_assembly_quality() {
let agent_id = AgentId::new();
let mut scored: Vec<ScoredMemory> = Vec::new();
for i in 0..2 {
let m = make_memory(
agent_id,
&format!(
"ANTI-PATTERN: Never use eval() in production code (pattern {})",
i
),
MemoryType::AntiPattern,
200 + i,
);
scored.push(ScoredMemory {
memory: m,
score: 0.7,
});
}
for i in 0..3 {
let m = make_memory(
agent_id,
&format!(
"CORRECTION: The API endpoint is /v2/users not /v1/users (correction {})",
i
),
MemoryType::Correction,
210 + i,
);
scored.push(ScoredMemory {
memory: m,
score: 0.75,
});
}
for i in 0..5 {
let mut m = make_memory(
agent_id,
&format!(
"Critical: User auth requires OAuth 2.0 PKCE flow (critical {})",
i
),
MemoryType::Semantic,
220 + i,
);
m.salience = 0.95;
scored.push(ScoredMemory {
memory: m,
score: 0.9,
});
}
for i in 0..10 {
let m = make_memory(
agent_id,
&format!(
"Project uses microservices architecture with Docker containers (detail {})",
i
),
MemoryType::Episodic,
230 + i,
);
scored.push(ScoredMemory {
memory: m,
score: 0.3 + (i as f32) * 0.04,
});
}
let structured_config = AssemblyConfig {
token_budget: 2000,
format: OutputFormat::Structured,
include_edges: false,
include_metadata: true,
};
let structured_ctx = ContextAssembler::assemble(scored.clone(), vec![], &structured_config);
assert!(structured_ctx.total_tokens > 0, "Should produce tokens");
assert!(
structured_ctx.total_tokens <= 2000,
"Should respect token budget (got {})",
structured_ctx.total_tokens
);
assert!(
!structured_ctx.blocks.is_empty(),
"Should have context blocks"
);
let compact_config = AssemblyConfig {
token_budget: 2000,
format: OutputFormat::Compact,
include_edges: false,
include_metadata: true,
};
let compact_ctx = ContextAssembler::assemble(scored.clone(), vec![], &compact_config);
assert!(
compact_ctx.total_tokens > 0,
"Compact should produce tokens"
);
if structured_ctx.metadata.included_count == compact_ctx.metadata.included_count {
assert!(
compact_ctx.total_tokens <= structured_ctx.total_tokens,
"Compact ({}) should use ≤ tokens than Structured ({})",
compact_ctx.total_tokens,
structured_ctx.total_tokens
);
} else {
assert!(
compact_ctx.metadata.included_count >= structured_ctx.metadata.included_count,
"Compact should fit more or equal memories"
);
}
let mut delta_tracker = DeltaTracker::new();
let delta_config = AssemblyConfig {
token_budget: 2000,
format: OutputFormat::Delta,
include_edges: false,
include_metadata: true,
};
let ctx1 =
ContextAssembler::assemble_delta(scored.clone(), vec![], &mut delta_tracker, &delta_config);
assert!(
ctx1.total_tokens > 0,
"First delta assembly should produce output"
);
let ctx2 =
ContextAssembler::assemble_delta(scored.clone(), vec![], &mut delta_tracker, &delta_config);
assert!(
ctx2.metadata.included_count == 0
|| ctx2.metadata.included_count <= ctx1.metadata.included_count,
"Second delta should include fewer or equal memories (delta-only): got {} vs first {}",
ctx2.metadata.included_count,
ctx1.metadata.included_count
);
let extra = make_memory(
agent_id,
"NEW: Just discovered a critical security vulnerability in auth module",
MemoryType::Correction,
250,
);
let mut scored_with_new = scored.clone();
scored_with_new.push(ScoredMemory {
memory: extra,
score: 0.99,
});
let ctx3 = ContextAssembler::assemble_delta(
scored_with_new,
vec![],
&mut delta_tracker,
&delta_config,
);
assert!(
ctx3.metadata.included_count >= 1,
"Third delta assembly should include the newly added memory"
);
}
#[test]
fn test_gdpr_forget() {
let dir = tempdir().unwrap();
let mut db = MenteDb::open(dir.path()).unwrap();
let agent_id = AgentId::new();
let mut memory_ids: Vec<MemoryId> = Vec::new();
let mut all_memories: Vec<MemoryNode> = Vec::new();
for i in 0..10 {
let m = make_memory(
agent_id,
&format!("User personal data entry number {i}: email, preferences, history"),
MemoryType::Episodic,
300 + i,
);
memory_ids.push(m.id);
all_memories.push(m.clone());
db.store(m).unwrap();
}
let mut edge_pairs: Vec<(MemoryId, MemoryId)> = Vec::new();
for i in 0..5 {
let edge = MemoryEdge {
source: memory_ids[i],
target: memory_ids[i + 1],
edge_type: EdgeType::Related,
weight: 0.7,
created_at: now_us(),
valid_from: None,
valid_until: None,
label: None,
};
edge_pairs.push((memory_ids[i], memory_ids[i + 1]));
db.relate(edge).unwrap();
}
let forget_engine = ForgetEngine::new();
let request = ForgetRequest {
agent_id: Some(agent_id),
space_id: None,
memory_ids: memory_ids.clone(),
reason: "GDPR Article 17 — Right to Erasure".to_string(),
requested_at: now_us(),
};
let plan = forget_engine.plan_forget(&request, &all_memories, &edge_pairs);
assert_eq!(
plan.deleted_memories, 10,
"Should plan to delete all 10 memories"
);
assert!(
plan.deleted_edges >= 5,
"Should plan to delete at least 5 edges, got {}",
plan.deleted_edges
);
let audit = forget_engine.generate_audit_log(&request, &plan);
assert!(!audit.is_empty(), "Audit log should be generated");
assert!(
audit.contains("GDPR")
|| audit.contains("erasure")
|| audit.contains("Erasure")
|| audit.contains("delete"),
"Audit log should reference the reason: {audit}"
);
for id in &memory_ids {
db.forget(*id).unwrap();
}
for i in 0..10u32 {
let emb = make_embedding(300 + i);
let results = db.recall_similar(&emb, 5).unwrap();
let found: Vec<_> = results
.iter()
.filter(|(id, _)| memory_ids.contains(id))
.collect();
assert!(
found.is_empty(),
"Memory {} should be gone after forget, but found {:?}",
i,
found
);
}
db.close().unwrap();
}