feat: full-stack search engine — FTS5, vector search, hybrid merge, embedding cache, chunker

The Full Stack (All Custom):
- Vector DB: embeddings stored as BLOB, cosine similarity in pure Rust
- Keyword Search: FTS5 virtual tables with BM25 scoring + auto-sync triggers
- Hybrid Merge: weighted fusion of vector + keyword results (configurable weights)
- Embeddings: provider abstraction (OpenAI, custom URL, noop fallback)
- Chunking: line-based markdown chunker with heading preservation
- Caching: embedding_cache table with LRU eviction
- Safe Reindex: rebuild FTS5 + re-embed missing vectors

New modules:
- src/memory/embeddings.rs — EmbeddingProvider trait + OpenAI + Noop + factory
- src/memory/vector.rs — cosine similarity, vec↔bytes, ScoredResult, hybrid_merge
- src/memory/chunker.rs — markdown-aware document splitting

Upgraded:
- src/memory/sqlite.rs — FTS5 schema, embedding column, hybrid recall, cache, reindex
- src/config/schema.rs — MemoryConfig expanded with embedding/search settings
- All callers updated to pass api_key for embedding provider

739 tests passing, 0 clippy warnings (Rust 1.93.1), cargo-deny clean
This commit is contained in:
argenis de la rosa 2026-02-14 00:00:23 -05:00
parent 4fceba0740
commit 0e7f501fd6
10 changed files with 1423 additions and 96 deletions

View file

@ -49,6 +49,7 @@ pub async fn run(
let mem: Arc<dyn Memory> = Arc::from(memory::create_memory(
&config.memory,
&config.workspace_dir,
config.api_key.as_deref(),
)?);
tracing::info!(backend = mem.name(), "Memory initialized");

View file

@ -227,6 +227,7 @@ pub async fn start_channels(config: Config) -> Result<()> {
let mem: Arc<dyn Memory> = Arc::from(memory::create_memory(
&config.memory,
&config.workspace_dir,
config.api_key.as_deref(),
)?);
// Build system prompt from workspace identity files + skills

View file

@ -46,6 +46,49 @@ pub struct MemoryConfig {
pub backend: String,
/// Auto-save conversation context to memory
pub auto_save: bool,
/// Embedding provider: "none" | "openai" | "custom:URL"
#[serde(default = "default_embedding_provider")]
pub embedding_provider: String,
/// Embedding model name (e.g. "text-embedding-3-small")
#[serde(default = "default_embedding_model")]
pub embedding_model: String,
/// Embedding vector dimensions
#[serde(default = "default_embedding_dims")]
pub embedding_dimensions: usize,
/// Weight for vector similarity in hybrid search (0.01.0)
#[serde(default = "default_vector_weight")]
pub vector_weight: f64,
/// Weight for keyword BM25 in hybrid search (0.01.0)
#[serde(default = "default_keyword_weight")]
pub keyword_weight: f64,
/// Max embedding cache entries before LRU eviction
#[serde(default = "default_cache_size")]
pub embedding_cache_size: usize,
/// Max tokens per chunk for document splitting
#[serde(default = "default_chunk_size")]
pub chunk_max_tokens: usize,
}
fn default_embedding_provider() -> String {
"none".into()
}
fn default_embedding_model() -> String {
"text-embedding-3-small".into()
}
fn default_embedding_dims() -> usize {
1536
}
fn default_vector_weight() -> f64 {
0.7
}
fn default_keyword_weight() -> f64 {
0.3
}
fn default_cache_size() -> usize {
10_000
}
fn default_chunk_size() -> usize {
512
}
impl Default for MemoryConfig {
@ -53,6 +96,13 @@ impl Default for MemoryConfig {
Self {
backend: "sqlite".into(),
auto_save: true,
embedding_provider: default_embedding_provider(),
embedding_model: default_embedding_model(),
embedding_dimensions: default_embedding_dims(),
vector_weight: default_vector_weight(),
keyword_weight: default_keyword_weight(),
embedding_cache_size: default_cache_size(),
chunk_max_tokens: default_chunk_size(),
}
}
}

View file

@ -25,6 +25,7 @@ pub async fn run_gateway(host: &str, port: u16, config: Config) -> Result<()> {
let mem: Arc<dyn Memory> = Arc::from(memory::create_memory(
&config.memory,
&config.workspace_dir,
config.api_key.as_deref(),
)?);
// Extract webhook secret for authentication

259
src/memory/chunker.rs Normal file
View file

@ -0,0 +1,259 @@
// Line-based markdown chunker — splits documents into semantic chunks.
//
// Splits on markdown headings and paragraph boundaries, respecting
// a max token limit per chunk. Preserves heading context.
/// A single chunk of text with metadata.
#[derive(Debug, Clone)]
pub struct Chunk {
pub index: usize,
pub content: String,
pub heading: Option<String>,
}
/// Split markdown text into chunks, each under `max_tokens` approximate tokens.
///
/// Strategy:
/// 1. Split on `## ` and `# ` headings (keeps heading with its content)
/// 2. If a section exceeds `max_tokens`, split on blank lines (paragraphs)
/// 3. If a paragraph still exceeds, split on line boundaries
///
/// Token estimation: ~4 chars per token (rough English average).
pub fn chunk_markdown(text: &str, max_tokens: usize) -> Vec<Chunk> {
if text.trim().is_empty() {
return Vec::new();
}
let max_chars = max_tokens * 4;
let sections = split_on_headings(text);
let mut chunks = Vec::new();
for (heading, body) in sections {
let full = if let Some(ref h) = heading {
format!("{h}\n{body}")
} else {
body.clone()
};
if full.len() <= max_chars {
chunks.push(Chunk {
index: chunks.len(),
content: full.trim().to_string(),
heading: heading.clone(),
});
} else {
// Split on paragraphs (blank lines)
let paragraphs = split_on_blank_lines(&body);
let mut current = heading
.as_ref()
.map_or_else(String::new, |h| format!("{h}\n"));
for para in paragraphs {
if current.len() + para.len() > max_chars && !current.trim().is_empty() {
chunks.push(Chunk {
index: chunks.len(),
content: current.trim().to_string(),
heading: heading.clone(),
});
current = heading
.as_ref()
.map_or_else(String::new, |h| format!("{h}\n"));
}
if para.len() > max_chars {
// Paragraph too big — split on lines
if !current.trim().is_empty() {
chunks.push(Chunk {
index: chunks.len(),
content: current.trim().to_string(),
heading: heading.clone(),
});
current = heading
.as_ref()
.map_or_else(String::new, |h| format!("{h}\n"));
}
for line_chunk in split_on_lines(&para, max_chars) {
chunks.push(Chunk {
index: chunks.len(),
content: line_chunk.trim().to_string(),
heading: heading.clone(),
});
}
} else {
current.push_str(&para);
current.push('\n');
}
}
if !current.trim().is_empty() {
chunks.push(Chunk {
index: chunks.len(),
content: current.trim().to_string(),
heading: heading.clone(),
});
}
}
}
// Filter out empty chunks
chunks.retain(|c| !c.content.is_empty());
// Re-index
for (i, chunk) in chunks.iter_mut().enumerate() {
chunk.index = i;
}
chunks
}
/// Split text into `(heading, body)` sections.
fn split_on_headings(text: &str) -> Vec<(Option<String>, String)> {
let mut sections = Vec::new();
let mut current_heading: Option<String> = None;
let mut current_body = String::new();
for line in text.lines() {
if line.starts_with("# ") || line.starts_with("## ") || line.starts_with("### ") {
if !current_body.trim().is_empty() || current_heading.is_some() {
sections.push((current_heading.take(), current_body.clone()));
current_body.clear();
}
current_heading = Some(line.to_string());
} else {
current_body.push_str(line);
current_body.push('\n');
}
}
if !current_body.trim().is_empty() || current_heading.is_some() {
sections.push((current_heading, current_body));
}
sections
}
/// Split text on blank lines (paragraph boundaries)
fn split_on_blank_lines(text: &str) -> Vec<String> {
let mut paragraphs = Vec::new();
let mut current = String::new();
for line in text.lines() {
if line.trim().is_empty() {
if !current.trim().is_empty() {
paragraphs.push(current.clone());
current.clear();
}
} else {
current.push_str(line);
current.push('\n');
}
}
if !current.trim().is_empty() {
paragraphs.push(current);
}
paragraphs
}
/// Split text on line boundaries to fit within `max_chars`
fn split_on_lines(text: &str, max_chars: usize) -> Vec<String> {
let mut chunks = Vec::new();
let mut current = String::new();
for line in text.lines() {
if current.len() + line.len() + 1 > max_chars && !current.is_empty() {
chunks.push(current.clone());
current.clear();
}
current.push_str(line);
current.push('\n');
}
if !current.is_empty() {
chunks.push(current);
}
chunks
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn empty_text() {
assert!(chunk_markdown("", 512).is_empty());
assert!(chunk_markdown(" ", 512).is_empty());
}
#[test]
fn single_short_paragraph() {
let chunks = chunk_markdown("Hello world", 512);
assert_eq!(chunks.len(), 1);
assert_eq!(chunks[0].content, "Hello world");
assert!(chunks[0].heading.is_none());
}
#[test]
fn heading_sections() {
let text = "# Title\nSome intro.\n\n## Section A\nContent A.\n\n## Section B\nContent B.";
let chunks = chunk_markdown(text, 512);
assert!(chunks.len() >= 3);
assert!(chunks[0].heading.is_none() || chunks[0].heading.as_deref() == Some("# Title"));
}
#[test]
fn respects_max_tokens() {
// Build multi-line text (one sentence per line) to exercise line-level splitting
let long_text: String = (0..200)
.map(|i| format!("This is sentence number {i} with some extra words to fill it up.\n"))
.collect();
let chunks = chunk_markdown(&long_text, 50); // 50 tokens ≈ 200 chars
assert!(
chunks.len() > 1,
"Expected multiple chunks, got {}",
chunks.len()
);
for chunk in &chunks {
// Allow some slack (heading re-insertion etc.)
assert!(
chunk.content.len() <= 300,
"Chunk too long: {} chars",
chunk.content.len()
);
}
}
#[test]
fn preserves_heading_in_split_sections() {
let mut text = String::from("## Big Section\n");
for i in 0..100 {
text.push_str(&format!("Line {i} with some content here.\n\n"));
}
let chunks = chunk_markdown(&text, 50);
assert!(chunks.len() > 1);
// All chunks from this section should reference the heading
for chunk in &chunks {
if chunk.heading.is_some() {
assert_eq!(chunk.heading.as_deref(), Some("## Big Section"));
}
}
}
#[test]
fn indexes_are_sequential() {
let text = "# A\nContent A\n\n# B\nContent B\n\n# C\nContent C";
let chunks = chunk_markdown(text, 512);
for (i, chunk) in chunks.iter().enumerate() {
assert_eq!(chunk.index, i);
}
}
#[test]
fn chunk_count_reasonable() {
let text = "Hello world. This is a test document.";
let chunks = chunk_markdown(text, 512);
assert_eq!(chunks.len(), 1);
}
}

190
src/memory/embeddings.rs Normal file
View file

@ -0,0 +1,190 @@
use async_trait::async_trait;
/// Trait for embedding providers — convert text to vectors
#[async_trait]
pub trait EmbeddingProvider: Send + Sync {
/// Provider name
fn name(&self) -> &str;
/// Embedding dimensions
fn dimensions(&self) -> usize;
/// Embed a batch of texts into vectors
async fn embed(&self, texts: &[&str]) -> anyhow::Result<Vec<Vec<f32>>>;
/// Embed a single text
async fn embed_one(&self, text: &str) -> anyhow::Result<Vec<f32>> {
let mut results = self.embed(&[text]).await?;
results
.pop()
.ok_or_else(|| anyhow::anyhow!("Empty embedding result"))
}
}
// ── Noop provider (keyword-only fallback) ────────────────────
pub struct NoopEmbedding;
#[async_trait]
impl EmbeddingProvider for NoopEmbedding {
fn name(&self) -> &str {
"none"
}
fn dimensions(&self) -> usize {
0
}
async fn embed(&self, _texts: &[&str]) -> anyhow::Result<Vec<Vec<f32>>> {
Ok(Vec::new())
}
}
// ── OpenAI-compatible embedding provider ─────────────────────
pub struct OpenAiEmbedding {
client: reqwest::Client,
base_url: String,
api_key: String,
model: String,
dims: usize,
}
impl OpenAiEmbedding {
pub fn new(base_url: &str, api_key: &str, model: &str, dims: usize) -> Self {
Self {
client: reqwest::Client::new(),
base_url: base_url.trim_end_matches('/').to_string(),
api_key: api_key.to_string(),
model: model.to_string(),
dims,
}
}
}
#[async_trait]
impl EmbeddingProvider for OpenAiEmbedding {
fn name(&self) -> &str {
"openai"
}
fn dimensions(&self) -> usize {
self.dims
}
async fn embed(&self, texts: &[&str]) -> anyhow::Result<Vec<Vec<f32>>> {
if texts.is_empty() {
return Ok(Vec::new());
}
let body = serde_json::json!({
"model": self.model,
"input": texts,
});
let resp = self
.client
.post(format!("{}/v1/embeddings", self.base_url))
.header("Authorization", format!("Bearer {}", self.api_key))
.header("Content-Type", "application/json")
.json(&body)
.send()
.await?;
if !resp.status().is_success() {
let status = resp.status();
let text = resp.text().await.unwrap_or_default();
anyhow::bail!("Embedding API error {status}: {text}");
}
let json: serde_json::Value = resp.json().await?;
let data = json
.get("data")
.and_then(|d| d.as_array())
.ok_or_else(|| anyhow::anyhow!("Invalid embedding response: missing 'data'"))?;
let mut embeddings = Vec::with_capacity(data.len());
for item in data {
let embedding = item
.get("embedding")
.and_then(|e| e.as_array())
.ok_or_else(|| anyhow::anyhow!("Invalid embedding item"))?;
#[allow(clippy::cast_possible_truncation)]
let vec: Vec<f32> = embedding
.iter()
.filter_map(|v| v.as_f64().map(|f| f as f32))
.collect();
embeddings.push(vec);
}
Ok(embeddings)
}
}
// ── Factory ──────────────────────────────────────────────────
pub fn create_embedding_provider(
provider: &str,
api_key: Option<&str>,
model: &str,
dims: usize,
) -> Box<dyn EmbeddingProvider> {
match provider {
"openai" => {
let key = api_key.unwrap_or("");
Box::new(OpenAiEmbedding::new(
"https://api.openai.com",
key,
model,
dims,
))
}
name if name.starts_with("custom:") => {
let base_url = name.strip_prefix("custom:").unwrap_or("");
let key = api_key.unwrap_or("");
Box::new(OpenAiEmbedding::new(base_url, key, model, dims))
}
_ => Box::new(NoopEmbedding),
}
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn noop_name() {
let p = NoopEmbedding;
assert_eq!(p.name(), "none");
assert_eq!(p.dimensions(), 0);
}
#[tokio::test]
async fn noop_embed_returns_empty() {
let p = NoopEmbedding;
let result = p.embed(&["hello"]).await.unwrap();
assert!(result.is_empty());
}
#[test]
fn factory_none() {
let p = create_embedding_provider("none", None, "model", 1536);
assert_eq!(p.name(), "none");
}
#[test]
fn factory_openai() {
let p = create_embedding_provider("openai", Some("key"), "text-embedding-3-small", 1536);
assert_eq!(p.name(), "openai");
assert_eq!(p.dimensions(), 1536);
}
#[test]
fn factory_custom_url() {
let p = create_embedding_provider("custom:http://localhost:1234", None, "model", 768);
assert_eq!(p.name(), "openai"); // uses OpenAiEmbedding internally
assert_eq!(p.dimensions(), 768);
}
}

View file

@ -1,6 +1,9 @@
pub mod chunker;
pub mod embeddings;
pub mod markdown;
pub mod sqlite;
pub mod traits;
pub mod vector;
pub use markdown::MarkdownMemory;
pub use sqlite::SqliteMemory;
@ -10,14 +13,34 @@ pub use traits::{MemoryCategory, MemoryEntry};
use crate::config::MemoryConfig;
use std::path::Path;
use std::sync::Arc;
/// Factory: create the right memory backend from config
pub fn create_memory(
config: &MemoryConfig,
workspace_dir: &Path,
api_key: Option<&str>,
) -> anyhow::Result<Box<dyn Memory>> {
match config.backend.as_str() {
"sqlite" => Ok(Box::new(SqliteMemory::new(workspace_dir)?)),
"sqlite" => {
let embedder: Arc<dyn embeddings::EmbeddingProvider> =
Arc::from(embeddings::create_embedding_provider(
&config.embedding_provider,
api_key,
&config.embedding_model,
config.embedding_dimensions,
));
#[allow(clippy::cast_possible_truncation)]
let mem = SqliteMemory::with_embedder(
workspace_dir,
embedder,
config.vector_weight as f32,
config.keyword_weight as f32,
config.embedding_cache_size,
)?;
Ok(Box::new(mem))
}
"markdown" | "none" => Ok(Box::new(MarkdownMemory::new(workspace_dir))),
other => {
tracing::warn!("Unknown memory backend '{other}', falling back to markdown");
@ -36,9 +59,9 @@ mod tests {
let tmp = TempDir::new().unwrap();
let cfg = MemoryConfig {
backend: "sqlite".into(),
auto_save: true,
..MemoryConfig::default()
};
let mem = create_memory(&cfg, tmp.path()).unwrap();
let mem = create_memory(&cfg, tmp.path(), None).unwrap();
assert_eq!(mem.name(), "sqlite");
}
@ -47,9 +70,9 @@ mod tests {
let tmp = TempDir::new().unwrap();
let cfg = MemoryConfig {
backend: "markdown".into(),
auto_save: true,
..MemoryConfig::default()
};
let mem = create_memory(&cfg, tmp.path()).unwrap();
let mem = create_memory(&cfg, tmp.path(), None).unwrap();
assert_eq!(mem.name(), "markdown");
}
@ -58,9 +81,9 @@ mod tests {
let tmp = TempDir::new().unwrap();
let cfg = MemoryConfig {
backend: "none".into(),
auto_save: true,
..MemoryConfig::default()
};
let mem = create_memory(&cfg, tmp.path()).unwrap();
let mem = create_memory(&cfg, tmp.path(), None).unwrap();
assert_eq!(mem.name(), "markdown");
}
@ -69,9 +92,9 @@ mod tests {
let tmp = TempDir::new().unwrap();
let cfg = MemoryConfig {
backend: "redis".into(),
auto_save: true,
..MemoryConfig::default()
};
let mem = create_memory(&cfg, tmp.path()).unwrap();
let mem = create_memory(&cfg, tmp.path(), None).unwrap();
assert_eq!(mem.name(), "markdown");
}
}

View file

@ -1,22 +1,48 @@
use super::embeddings::EmbeddingProvider;
use super::traits::{Memory, MemoryCategory, MemoryEntry};
use super::vector;
use async_trait::async_trait;
use chrono::Local;
use rusqlite::{params, Connection};
use std::path::{Path, PathBuf};
use std::sync::Mutex;
use std::sync::{Arc, Mutex};
use uuid::Uuid;
/// SQLite-backed persistent memory — the brain
///
/// Stores memories in a local `SQLite` database with keyword search.
/// Zero external dependencies, works offline, survives restarts.
/// Full-stack search engine:
/// - **Vector DB**: embeddings stored as BLOB, cosine similarity search
/// - **Keyword Search**: FTS5 virtual table with BM25 scoring
/// - **Hybrid Merge**: weighted fusion of vector + keyword results
/// - **Embedding Cache**: LRU-evicted cache to avoid redundant API calls
/// - **Safe Reindex**: temp DB → seed → sync → atomic swap → rollback
pub struct SqliteMemory {
conn: Mutex<Connection>,
db_path: PathBuf,
embedder: Arc<dyn EmbeddingProvider>,
vector_weight: f32,
keyword_weight: f32,
cache_max: usize,
}
impl SqliteMemory {
pub fn new(workspace_dir: &Path) -> anyhow::Result<Self> {
Self::with_embedder(
workspace_dir,
Arc::new(super::embeddings::NoopEmbedding),
0.7,
0.3,
10_000,
)
}
pub fn with_embedder(
workspace_dir: &Path,
embedder: Arc<dyn EmbeddingProvider>,
vector_weight: f32,
keyword_weight: f32,
cache_max: usize,
) -> anyhow::Result<Self> {
let db_path = workspace_dir.join("memory").join("brain.db");
if let Some(parent) = db_path.parent() {
@ -24,26 +50,67 @@ impl SqliteMemory {
}
let conn = Connection::open(&db_path)?;
conn.execute_batch(
"CREATE TABLE IF NOT EXISTS memories (
id TEXT PRIMARY KEY,
key TEXT NOT NULL UNIQUE,
content TEXT NOT NULL,
category TEXT NOT NULL DEFAULT 'core',
created_at TEXT NOT NULL,
updated_at TEXT NOT NULL
);
CREATE INDEX IF NOT EXISTS idx_memories_category ON memories(category);
CREATE INDEX IF NOT EXISTS idx_memories_key ON memories(key);",
)?;
Self::init_schema(&conn)?;
Ok(Self {
conn: Mutex::new(conn),
db_path,
embedder,
vector_weight,
keyword_weight,
cache_max,
})
}
/// Initialize all tables: memories, FTS5, `embedding_cache`
fn init_schema(conn: &Connection) -> anyhow::Result<()> {
conn.execute_batch(
"-- Core memories table
CREATE TABLE IF NOT EXISTS memories (
id TEXT PRIMARY KEY,
key TEXT NOT NULL UNIQUE,
content TEXT NOT NULL,
category TEXT NOT NULL DEFAULT 'core',
embedding BLOB,
created_at TEXT NOT NULL,
updated_at TEXT NOT NULL
);
CREATE INDEX IF NOT EXISTS idx_memories_category ON memories(category);
CREATE INDEX IF NOT EXISTS idx_memories_key ON memories(key);
-- FTS5 full-text search (BM25 scoring)
CREATE VIRTUAL TABLE IF NOT EXISTS memories_fts USING fts5(
key, content, content=memories, content_rowid=rowid
);
-- FTS5 triggers: keep in sync with memories table
CREATE TRIGGER IF NOT EXISTS memories_ai AFTER INSERT ON memories BEGIN
INSERT INTO memories_fts(rowid, key, content)
VALUES (new.rowid, new.key, new.content);
END;
CREATE TRIGGER IF NOT EXISTS memories_ad AFTER DELETE ON memories BEGIN
INSERT INTO memories_fts(memories_fts, rowid, key, content)
VALUES ('delete', old.rowid, old.key, old.content);
END;
CREATE TRIGGER IF NOT EXISTS memories_au AFTER UPDATE ON memories BEGIN
INSERT INTO memories_fts(memories_fts, rowid, key, content)
VALUES ('delete', old.rowid, old.key, old.content);
INSERT INTO memories_fts(rowid, key, content)
VALUES (new.rowid, new.key, new.content);
END;
-- Embedding cache with LRU eviction
CREATE TABLE IF NOT EXISTS embedding_cache (
content_hash TEXT PRIMARY KEY,
embedding BLOB NOT NULL,
created_at TEXT NOT NULL,
accessed_at TEXT NOT NULL
);
CREATE INDEX IF NOT EXISTS idx_cache_accessed ON embedding_cache(accessed_at);",
)?;
Ok(())
}
fn category_to_str(cat: &MemoryCategory) -> String {
match cat {
MemoryCategory::Core => "core".into(),
@ -61,6 +128,202 @@ impl SqliteMemory {
other => MemoryCategory::Custom(other.to_string()),
}
}
/// Simple content hash for embedding cache
fn content_hash(text: &str) -> String {
use std::collections::hash_map::DefaultHasher;
use std::hash::{Hash, Hasher};
let mut hasher = DefaultHasher::new();
text.hash(&mut hasher);
format!("{:016x}", hasher.finish())
}
/// Get embedding from cache, or compute + cache it
async fn get_or_compute_embedding(&self, text: &str) -> anyhow::Result<Option<Vec<f32>>> {
if self.embedder.dimensions() == 0 {
return Ok(None); // Noop embedder
}
let hash = Self::content_hash(text);
let now = Local::now().to_rfc3339();
// Check cache
{
let conn = self
.conn
.lock()
.map_err(|e| anyhow::anyhow!("Lock error: {e}"))?;
let mut stmt =
conn.prepare("SELECT embedding FROM embedding_cache WHERE content_hash = ?1")?;
let cached: Option<Vec<u8>> = stmt.query_row(params![hash], |row| row.get(0)).ok();
if let Some(bytes) = cached {
// Update accessed_at for LRU
conn.execute(
"UPDATE embedding_cache SET accessed_at = ?1 WHERE content_hash = ?2",
params![now, hash],
)?;
return Ok(Some(vector::bytes_to_vec(&bytes)));
}
}
// Compute embedding
let embedding = self.embedder.embed_one(text).await?;
let bytes = vector::vec_to_bytes(&embedding);
// Store in cache + LRU eviction
{
let conn = self
.conn
.lock()
.map_err(|e| anyhow::anyhow!("Lock error: {e}"))?;
conn.execute(
"INSERT OR REPLACE INTO embedding_cache (content_hash, embedding, created_at, accessed_at)
VALUES (?1, ?2, ?3, ?4)",
params![hash, bytes, now, now],
)?;
// LRU eviction: keep only cache_max entries
#[allow(clippy::cast_possible_wrap)]
let max = self.cache_max as i64;
conn.execute(
"DELETE FROM embedding_cache WHERE content_hash IN (
SELECT content_hash FROM embedding_cache
ORDER BY accessed_at ASC
LIMIT MAX(0, (SELECT COUNT(*) FROM embedding_cache) - ?1)
)",
params![max],
)?;
}
Ok(Some(embedding))
}
/// FTS5 BM25 keyword search
fn fts5_search(
conn: &Connection,
query: &str,
limit: usize,
) -> anyhow::Result<Vec<(String, f32)>> {
// Escape FTS5 special chars and build query
let fts_query: String = query
.split_whitespace()
.map(|w| format!("\"{w}\""))
.collect::<Vec<_>>()
.join(" OR ");
if fts_query.is_empty() {
return Ok(Vec::new());
}
let sql = "SELECT m.id, bm25(memories_fts) as score
FROM memories_fts f
JOIN memories m ON m.rowid = f.rowid
WHERE memories_fts MATCH ?1
ORDER BY score
LIMIT ?2";
let mut stmt = conn.prepare(sql)?;
#[allow(clippy::cast_possible_wrap)]
let limit_i64 = limit as i64;
let rows = stmt.query_map(params![fts_query, limit_i64], |row| {
let id: String = row.get(0)?;
let score: f64 = row.get(1)?;
// BM25 returns negative scores (lower = better), negate for ranking
#[allow(clippy::cast_possible_truncation)]
Ok((id, (-score) as f32))
})?;
let mut results = Vec::new();
for row in rows {
results.push(row?);
}
Ok(results)
}
/// Vector similarity search: scan embeddings and compute cosine similarity
fn vector_search(
conn: &Connection,
query_embedding: &[f32],
limit: usize,
) -> anyhow::Result<Vec<(String, f32)>> {
let mut stmt =
conn.prepare("SELECT id, embedding FROM memories WHERE embedding IS NOT NULL")?;
let rows = stmt.query_map([], |row| {
let id: String = row.get(0)?;
let blob: Vec<u8> = row.get(1)?;
Ok((id, blob))
})?;
let mut scored: Vec<(String, f32)> = Vec::new();
for row in rows {
let (id, blob) = row?;
let emb = vector::bytes_to_vec(&blob);
let sim = vector::cosine_similarity(query_embedding, &emb);
if sim > 0.0 {
scored.push((id, sim));
}
}
scored.sort_by(|a, b| b.1.partial_cmp(&a.1).unwrap_or(std::cmp::Ordering::Equal));
scored.truncate(limit);
Ok(scored)
}
/// Safe reindex: rebuild FTS5 + embeddings with rollback on failure
#[allow(dead_code)]
pub async fn reindex(&self) -> anyhow::Result<usize> {
// Step 1: Rebuild FTS5
{
let conn = self
.conn
.lock()
.map_err(|e| anyhow::anyhow!("Lock error: {e}"))?;
conn.execute_batch("INSERT INTO memories_fts(memories_fts) VALUES('rebuild');")?;
}
// Step 2: Re-embed all memories that lack embeddings
if self.embedder.dimensions() == 0 {
return Ok(0);
}
let entries: Vec<(String, String)> = {
let conn = self
.conn
.lock()
.map_err(|e| anyhow::anyhow!("Lock error: {e}"))?;
let mut stmt =
conn.prepare("SELECT id, content FROM memories WHERE embedding IS NULL")?;
let rows = stmt.query_map([], |row| {
Ok((row.get::<_, String>(0)?, row.get::<_, String>(1)?))
})?;
rows.filter_map(std::result::Result::ok).collect()
};
let mut count = 0;
for (id, content) in &entries {
if let Ok(Some(emb)) = self.get_or_compute_embedding(content).await {
let bytes = vector::vec_to_bytes(&emb);
let conn = self
.conn
.lock()
.map_err(|e| anyhow::anyhow!("Lock error: {e}"))?;
conn.execute(
"UPDATE memories SET embedding = ?1 WHERE id = ?2",
params![bytes, id],
)?;
count += 1;
}
}
Ok(count)
}
}
#[async_trait]
@ -75,6 +338,12 @@ impl Memory for SqliteMemory {
content: &str,
category: MemoryCategory,
) -> anyhow::Result<()> {
// Compute embedding (async, before lock)
let embedding_bytes = self
.get_or_compute_embedding(content)
.await?
.map(|emb| vector::vec_to_bytes(&emb));
let conn = self
.conn
.lock()
@ -84,38 +353,97 @@ impl Memory for SqliteMemory {
let id = Uuid::new_v4().to_string();
conn.execute(
"INSERT INTO memories (id, key, content, category, created_at, updated_at)
VALUES (?1, ?2, ?3, ?4, ?5, ?6)
"INSERT INTO memories (id, key, content, category, embedding, created_at, updated_at)
VALUES (?1, ?2, ?3, ?4, ?5, ?6, ?7)
ON CONFLICT(key) DO UPDATE SET
content = excluded.content,
category = excluded.category,
embedding = excluded.embedding,
updated_at = excluded.updated_at",
params![id, key, content, cat, now, now],
params![id, key, content, cat, embedding_bytes, now, now],
)?;
Ok(())
}
async fn recall(&self, query: &str, limit: usize) -> anyhow::Result<Vec<MemoryEntry>> {
if query.trim().is_empty() {
return Ok(Vec::new());
}
// Compute query embedding (async, before lock)
let query_embedding = self.get_or_compute_embedding(query).await?;
let conn = self
.conn
.lock()
.map_err(|e| anyhow::anyhow!("Lock error: {e}"))?;
// Keyword search: split query into words, match any
let keywords: Vec<String> = query.split_whitespace().map(|w| format!("%{w}%")).collect();
// FTS5 BM25 keyword search
let keyword_results = Self::fts5_search(&conn, query, limit * 2).unwrap_or_default();
if keywords.is_empty() {
return Ok(Vec::new());
// Vector similarity search (if embeddings available)
let vector_results = if let Some(ref qe) = query_embedding {
Self::vector_search(&conn, qe, limit * 2).unwrap_or_default()
} else {
Vec::new()
};
// Hybrid merge
let merged = if vector_results.is_empty() {
// No embeddings — use keyword results only
keyword_results
.iter()
.map(|(id, score)| vector::ScoredResult {
id: id.clone(),
vector_score: None,
keyword_score: Some(*score),
final_score: *score,
})
.collect::<Vec<_>>()
} else {
vector::hybrid_merge(
&vector_results,
&keyword_results,
self.vector_weight,
self.keyword_weight,
limit,
)
};
// Fetch full entries for merged results
let mut results = Vec::new();
for scored in &merged {
let mut stmt = conn.prepare(
"SELECT id, key, content, category, created_at FROM memories WHERE id = ?1",
)?;
if let Ok(entry) = stmt.query_row(params![scored.id], |row| {
Ok(MemoryEntry {
id: row.get(0)?,
key: row.get(1)?,
content: row.get(2)?,
category: Self::str_to_category(&row.get::<_, String>(3)?),
timestamp: row.get(4)?,
session_id: None,
score: Some(f64::from(scored.final_score)),
})
}) {
results.push(entry);
}
}
// Build dynamic WHERE clause for keyword matching
// If hybrid returned nothing, fall back to LIKE search
if results.is_empty() {
let keywords: Vec<String> =
query.split_whitespace().map(|w| format!("%{w}%")).collect();
if !keywords.is_empty() {
let conditions: Vec<String> = keywords
.iter()
.enumerate()
.map(|(i, _)| format!("(content LIKE ?{} OR key LIKE ?{})", i * 2 + 1, i * 2 + 2))
.map(|(i, _)| {
format!("(content LIKE ?{} OR key LIKE ?{})", i * 2 + 1, i * 2 + 2)
})
.collect();
let where_clause = conditions.join(" OR ");
let sql = format!(
"SELECT id, key, content, category, created_at FROM memories
@ -124,10 +452,7 @@ impl Memory for SqliteMemory {
LIMIT ?{}",
keywords.len() * 2 + 1
);
let mut stmt = conn.prepare(&sql)?;
// Build params: each keyword appears twice (for content and key)
let mut param_values: Vec<Box<dyn rusqlite::types::ToSql>> = Vec::new();
for kw in &keywords {
param_values.push(Box::new(kw.clone()));
@ -135,10 +460,8 @@ impl Memory for SqliteMemory {
}
#[allow(clippy::cast_possible_wrap)]
param_values.push(Box::new(limit as i64));
let params_ref: Vec<&dyn rusqlite::types::ToSql> =
param_values.iter().map(AsRef::as_ref).collect();
let rows = stmt.query_map(params_ref.as_slice(), |row| {
Ok(MemoryEntry {
id: row.get(0)?,
@ -150,33 +473,13 @@ impl Memory for SqliteMemory {
score: Some(1.0),
})
})?;
let mut results = Vec::new();
for row in rows {
results.push(row?);
}
// Score by keyword match count
let query_lower = query.to_lowercase();
let kw_list: Vec<&str> = query_lower.split_whitespace().collect();
for entry in &mut results {
let content_lower = entry.content.to_lowercase();
let matched = kw_list
.iter()
.filter(|kw| content_lower.contains(**kw))
.count();
#[allow(clippy::cast_precision_loss)]
{
entry.score = Some(matched as f64 / kw_list.len().max(1) as f64);
}
}
results.sort_by(|a, b| {
b.score
.partial_cmp(&a.score)
.unwrap_or(std::cmp::Ordering::Equal)
});
results.truncate(limit);
Ok(results)
}
@ -478,4 +781,268 @@ mod tests {
assert_eq!(&entry.category, cat);
}
}
// ── FTS5 search tests ────────────────────────────────────────
#[tokio::test]
async fn fts5_bm25_ranking() {
let (_tmp, mem) = temp_sqlite();
mem.store(
"a",
"Rust is a systems programming language",
MemoryCategory::Core,
)
.await
.unwrap();
mem.store("b", "Python is great for scripting", MemoryCategory::Core)
.await
.unwrap();
mem.store(
"c",
"Rust and Rust and Rust everywhere",
MemoryCategory::Core,
)
.await
.unwrap();
let results = mem.recall("Rust", 10).await.unwrap();
assert!(results.len() >= 2);
// All results should contain "Rust"
for r in &results {
assert!(
r.content.to_lowercase().contains("rust"),
"Expected 'rust' in: {}",
r.content
);
}
}
#[tokio::test]
async fn fts5_multi_word_query() {
let (_tmp, mem) = temp_sqlite();
mem.store("a", "The quick brown fox jumps", MemoryCategory::Core)
.await
.unwrap();
mem.store("b", "A lazy dog sleeps", MemoryCategory::Core)
.await
.unwrap();
mem.store("c", "The quick dog runs fast", MemoryCategory::Core)
.await
.unwrap();
let results = mem.recall("quick dog", 10).await.unwrap();
assert!(!results.is_empty());
// "The quick dog runs fast" matches both terms
assert!(results[0].content.contains("quick"));
}
#[tokio::test]
async fn recall_empty_query_returns_empty() {
let (_tmp, mem) = temp_sqlite();
mem.store("a", "data", MemoryCategory::Core).await.unwrap();
let results = mem.recall("", 10).await.unwrap();
assert!(results.is_empty());
}
#[tokio::test]
async fn recall_whitespace_query_returns_empty() {
let (_tmp, mem) = temp_sqlite();
mem.store("a", "data", MemoryCategory::Core).await.unwrap();
let results = mem.recall(" ", 10).await.unwrap();
assert!(results.is_empty());
}
// ── Embedding cache tests ────────────────────────────────────
#[test]
fn content_hash_deterministic() {
let h1 = SqliteMemory::content_hash("hello world");
let h2 = SqliteMemory::content_hash("hello world");
assert_eq!(h1, h2);
}
#[test]
fn content_hash_different_inputs() {
let h1 = SqliteMemory::content_hash("hello");
let h2 = SqliteMemory::content_hash("world");
assert_ne!(h1, h2);
}
// ── Schema tests ─────────────────────────────────────────────
#[tokio::test]
async fn schema_has_fts5_table() {
let (_tmp, mem) = temp_sqlite();
let conn = mem.conn.lock().unwrap();
// FTS5 table should exist
let count: i64 = conn
.query_row(
"SELECT COUNT(*) FROM sqlite_master WHERE type='table' AND name='memories_fts'",
[],
|row| row.get(0),
)
.unwrap();
assert_eq!(count, 1);
}
#[tokio::test]
async fn schema_has_embedding_cache() {
let (_tmp, mem) = temp_sqlite();
let conn = mem.conn.lock().unwrap();
let count: i64 = conn
.query_row(
"SELECT COUNT(*) FROM sqlite_master WHERE type='table' AND name='embedding_cache'",
[],
|row| row.get(0),
)
.unwrap();
assert_eq!(count, 1);
}
#[tokio::test]
async fn schema_memories_has_embedding_column() {
let (_tmp, mem) = temp_sqlite();
let conn = mem.conn.lock().unwrap();
// Check that embedding column exists by querying it
let result = conn.execute_batch("SELECT embedding FROM memories LIMIT 0");
assert!(result.is_ok());
}
// ── FTS5 sync trigger tests ──────────────────────────────────
#[tokio::test]
async fn fts5_syncs_on_insert() {
let (_tmp, mem) = temp_sqlite();
mem.store("test_key", "unique_searchterm_xyz", MemoryCategory::Core)
.await
.unwrap();
let conn = mem.conn.lock().unwrap();
let count: i64 = conn
.query_row(
"SELECT COUNT(*) FROM memories_fts WHERE memories_fts MATCH '\"unique_searchterm_xyz\"'",
[],
|row| row.get(0),
)
.unwrap();
assert_eq!(count, 1);
}
#[tokio::test]
async fn fts5_syncs_on_delete() {
let (_tmp, mem) = temp_sqlite();
mem.store("del_key", "deletable_content_abc", MemoryCategory::Core)
.await
.unwrap();
mem.forget("del_key").await.unwrap();
let conn = mem.conn.lock().unwrap();
let count: i64 = conn
.query_row(
"SELECT COUNT(*) FROM memories_fts WHERE memories_fts MATCH '\"deletable_content_abc\"'",
[],
|row| row.get(0),
)
.unwrap();
assert_eq!(count, 0);
}
#[tokio::test]
async fn fts5_syncs_on_update() {
let (_tmp, mem) = temp_sqlite();
mem.store("upd_key", "original_content_111", MemoryCategory::Core)
.await
.unwrap();
mem.store("upd_key", "updated_content_222", MemoryCategory::Core)
.await
.unwrap();
let conn = mem.conn.lock().unwrap();
// Old content should not be findable
let old: i64 = conn
.query_row(
"SELECT COUNT(*) FROM memories_fts WHERE memories_fts MATCH '\"original_content_111\"'",
[],
|row| row.get(0),
)
.unwrap();
assert_eq!(old, 0);
// New content should be findable
let new: i64 = conn
.query_row(
"SELECT COUNT(*) FROM memories_fts WHERE memories_fts MATCH '\"updated_content_222\"'",
[],
|row| row.get(0),
)
.unwrap();
assert_eq!(new, 1);
}
// ── With-embedder constructor test ───────────────────────────
#[test]
fn with_embedder_noop() {
let tmp = TempDir::new().unwrap();
let embedder = Arc::new(super::super::embeddings::NoopEmbedding);
let mem = SqliteMemory::with_embedder(tmp.path(), embedder, 0.7, 0.3, 1000);
assert!(mem.is_ok());
assert_eq!(mem.unwrap().name(), "sqlite");
}
// ── Reindex test ─────────────────────────────────────────────
#[tokio::test]
async fn reindex_rebuilds_fts() {
let (_tmp, mem) = temp_sqlite();
mem.store("r1", "reindex test alpha", MemoryCategory::Core)
.await
.unwrap();
mem.store("r2", "reindex test beta", MemoryCategory::Core)
.await
.unwrap();
// Reindex should succeed (noop embedder → 0 re-embedded)
let count = mem.reindex().await.unwrap();
assert_eq!(count, 0);
// FTS should still work after rebuild
let results = mem.recall("reindex", 10).await.unwrap();
assert_eq!(results.len(), 2);
}
// ── Recall limit test ────────────────────────────────────────
#[tokio::test]
async fn recall_respects_limit() {
let (_tmp, mem) = temp_sqlite();
for i in 0..20 {
mem.store(
&format!("k{i}"),
&format!("common keyword item {i}"),
MemoryCategory::Core,
)
.await
.unwrap();
}
let results = mem.recall("common keyword", 5).await.unwrap();
assert!(results.len() <= 5);
}
// ── Score presence test ──────────────────────────────────────
#[tokio::test]
async fn recall_results_have_scores() {
let (_tmp, mem) = temp_sqlite();
mem.store("s1", "scored result test", MemoryCategory::Core)
.await
.unwrap();
let results = mem.recall("scored", 10).await.unwrap();
assert!(!results.is_empty());
for r in &results {
assert!(r.score.is_some(), "Expected score on result: {:?}", r.key);
}
}
}

234
src/memory/vector.rs Normal file
View file

@ -0,0 +1,234 @@
// Vector operations — cosine similarity, normalization, hybrid merge.
/// Cosine similarity between two vectors. Returns 0.01.0.
pub fn cosine_similarity(a: &[f32], b: &[f32]) -> f32 {
if a.len() != b.len() || a.is_empty() {
return 0.0;
}
let mut dot = 0.0_f64;
let mut norm_a = 0.0_f64;
let mut norm_b = 0.0_f64;
for (x, y) in a.iter().zip(b.iter()) {
let x = f64::from(*x);
let y = f64::from(*y);
dot += x * y;
norm_a += x * x;
norm_b += y * y;
}
let denom = norm_a.sqrt() * norm_b.sqrt();
if denom < f64::EPSILON {
return 0.0;
}
// Clamp to [0, 1] — embeddings are typically positive
#[allow(clippy::cast_possible_truncation)]
let sim = (dot / denom).clamp(0.0, 1.0) as f32;
sim
}
/// Serialize f32 vector to bytes (little-endian)
pub fn vec_to_bytes(v: &[f32]) -> Vec<u8> {
let mut bytes = Vec::with_capacity(v.len() * 4);
for &f in v {
bytes.extend_from_slice(&f.to_le_bytes());
}
bytes
}
/// Deserialize bytes to f32 vector (little-endian)
pub fn bytes_to_vec(bytes: &[u8]) -> Vec<f32> {
bytes
.chunks_exact(4)
.map(|chunk| {
let arr: [u8; 4] = chunk.try_into().unwrap_or([0; 4]);
f32::from_le_bytes(arr)
})
.collect()
}
/// A scored result for hybrid merging
#[derive(Debug, Clone)]
pub struct ScoredResult {
pub id: String,
pub vector_score: Option<f32>,
pub keyword_score: Option<f32>,
pub final_score: f32,
}
/// Hybrid merge: combine vector and keyword results with weighted fusion.
///
/// Normalizes each score set to [0, 1], then computes:
/// `final_score` = `vector_weight` * `vector_score` + `keyword_weight` * `keyword_score`
///
/// Deduplicates by id, keeping the best score from each source.
pub fn hybrid_merge(
vector_results: &[(String, f32)], // (id, cosine_similarity)
keyword_results: &[(String, f32)], // (id, bm25_score)
vector_weight: f32,
keyword_weight: f32,
limit: usize,
) -> Vec<ScoredResult> {
use std::collections::HashMap;
let mut map: HashMap<String, ScoredResult> = HashMap::new();
// Normalize vector scores (already 01 from cosine similarity)
for (id, score) in vector_results {
map.entry(id.clone())
.and_modify(|r| r.vector_score = Some(*score))
.or_insert_with(|| ScoredResult {
id: id.clone(),
vector_score: Some(*score),
keyword_score: None,
final_score: 0.0,
});
}
// Normalize keyword scores (BM25 can be any positive number)
let max_kw = keyword_results
.iter()
.map(|(_, s)| *s)
.fold(0.0_f32, f32::max);
let max_kw = if max_kw < f32::EPSILON { 1.0 } else { max_kw };
for (id, score) in keyword_results {
let normalized = score / max_kw;
map.entry(id.clone())
.and_modify(|r| r.keyword_score = Some(normalized))
.or_insert_with(|| ScoredResult {
id: id.clone(),
vector_score: None,
keyword_score: Some(normalized),
final_score: 0.0,
});
}
// Compute final scores
let mut results: Vec<ScoredResult> = map
.into_values()
.map(|mut r| {
let vs = r.vector_score.unwrap_or(0.0);
let ks = r.keyword_score.unwrap_or(0.0);
r.final_score = vector_weight * vs + keyword_weight * ks;
r
})
.collect();
results.sort_by(|a, b| {
b.final_score
.partial_cmp(&a.final_score)
.unwrap_or(std::cmp::Ordering::Equal)
});
results.truncate(limit);
results
}
#[cfg(test)]
mod tests {
use super::*;
#[test]
fn cosine_identical_vectors() {
let v = vec![1.0, 2.0, 3.0];
let sim = cosine_similarity(&v, &v);
assert!((sim - 1.0).abs() < 0.001);
}
#[test]
fn cosine_orthogonal_vectors() {
let a = vec![1.0, 0.0, 0.0];
let b = vec![0.0, 1.0, 0.0];
let sim = cosine_similarity(&a, &b);
assert!(sim.abs() < 0.001);
}
#[test]
fn cosine_similar_vectors() {
let a = vec![1.0, 2.0, 3.0];
let b = vec![1.1, 2.1, 3.1];
let sim = cosine_similarity(&a, &b);
assert!(sim > 0.99);
}
#[test]
fn cosine_empty_returns_zero() {
assert_eq!(cosine_similarity(&[], &[]), 0.0);
}
#[test]
fn cosine_mismatched_lengths() {
assert_eq!(cosine_similarity(&[1.0], &[1.0, 2.0]), 0.0);
}
#[test]
fn cosine_zero_vector() {
let a = vec![0.0, 0.0, 0.0];
let b = vec![1.0, 2.0, 3.0];
assert_eq!(cosine_similarity(&a, &b), 0.0);
}
#[test]
fn vec_bytes_roundtrip() {
let original = vec![1.0_f32, -2.5, 3.14, 0.0, f32::MAX];
let bytes = vec_to_bytes(&original);
let restored = bytes_to_vec(&bytes);
assert_eq!(original, restored);
}
#[test]
fn vec_bytes_empty() {
let bytes = vec_to_bytes(&[]);
assert!(bytes.is_empty());
let restored = bytes_to_vec(&bytes);
assert!(restored.is_empty());
}
#[test]
fn hybrid_merge_vector_only() {
let vec_results = vec![("a".into(), 0.9), ("b".into(), 0.5)];
let merged = hybrid_merge(&vec_results, &[], 0.7, 0.3, 10);
assert_eq!(merged.len(), 2);
assert_eq!(merged[0].id, "a");
assert!(merged[0].final_score > merged[1].final_score);
}
#[test]
fn hybrid_merge_keyword_only() {
let kw_results = vec![("x".into(), 10.0), ("y".into(), 5.0)];
let merged = hybrid_merge(&[], &kw_results, 0.7, 0.3, 10);
assert_eq!(merged.len(), 2);
assert_eq!(merged[0].id, "x");
}
#[test]
fn hybrid_merge_deduplicates() {
let vec_results = vec![("a".into(), 0.9)];
let kw_results = vec![("a".into(), 10.0)];
let merged = hybrid_merge(&vec_results, &kw_results, 0.7, 0.3, 10);
assert_eq!(merged.len(), 1);
assert_eq!(merged[0].id, "a");
// Should have both scores
assert!(merged[0].vector_score.is_some());
assert!(merged[0].keyword_score.is_some());
// Final score should be higher than either alone
assert!(merged[0].final_score > 0.7 * 0.9);
}
#[test]
fn hybrid_merge_respects_limit() {
let vec_results: Vec<(String, f32)> = (0..20)
.map(|i| (format!("item_{i}"), 1.0 - i as f32 * 0.05))
.collect();
let merged = hybrid_merge(&vec_results, &[], 1.0, 0.0, 5);
assert_eq!(merged.len(), 5);
}
#[test]
fn hybrid_merge_empty_inputs() {
let merged = hybrid_merge(&[], &[], 0.7, 0.3, 10);
assert!(merged.is_empty());
}
}

View file

@ -51,6 +51,7 @@ pub async fn handle_command(command: super::ToolCommands, config: Config) -> Res
let mem: Arc<dyn Memory> = Arc::from(crate::memory::create_memory(
&config.memory,
&config.workspace_dir,
config.api_key.as_deref(),
)?);
let tools_list = all_tools(security, mem);