3e770df332
ObservationCompiler.queryObservationsMulti and querySummariesMulti WHERE clause extended with OR merged_into_project IN (...), so a parent-project read pulls in rows originally written under any child worktree's composite name once merged. SearchManager wraps the Chroma project filter in \$or so semantic search behaves identically. ChromaSync baseMetadata now carries merged_into_project on new embeddings; existing rows are patched retroactively by the adoption engine. Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
916 lines
30 KiB
TypeScript
916 lines
30 KiB
TypeScript
/**
|
|
* ChromaSync Service
|
|
*
|
|
* Automatically syncs observations and session summaries to ChromaDB via MCP.
|
|
* This service provides real-time semantic search capabilities by maintaining
|
|
* a vector database synchronized with SQLite.
|
|
*
|
|
* Uses ChromaMcpManager to communicate with chroma-mcp over stdio MCP protocol.
|
|
* The chroma-mcp server handles its own embedding and persistent storage,
|
|
* eliminating the need for chromadb npm package and ONNX/WASM dependencies.
|
|
*
|
|
* Design: Fail-fast with no fallbacks - if Chroma is unavailable, syncing fails.
|
|
*/
|
|
|
|
import { ChromaMcpManager } from './ChromaMcpManager.js';
|
|
import { ParsedObservation, ParsedSummary } from '../../sdk/parser.js';
|
|
import { SessionStore } from '../sqlite/SessionStore.js';
|
|
import { logger } from '../../utils/logger.js';
|
|
import { parseFileList } from '../sqlite/observations/files.js';
|
|
|
|
interface ChromaDocument {
|
|
id: string;
|
|
document: string;
|
|
metadata: Record<string, string | number>;
|
|
}
|
|
|
|
interface StoredObservation {
|
|
id: number;
|
|
memory_session_id: string;
|
|
project: string;
|
|
merged_into_project: string | null;
|
|
text: string | null;
|
|
type: string;
|
|
title: string | null;
|
|
subtitle: string | null;
|
|
facts: string | null; // JSON
|
|
narrative: string | null;
|
|
concepts: string | null; // JSON
|
|
files_read: string | null; // JSON
|
|
files_modified: string | null; // JSON
|
|
prompt_number: number;
|
|
discovery_tokens: number; // ROI metrics
|
|
created_at: string;
|
|
created_at_epoch: number;
|
|
}
|
|
|
|
interface StoredSummary {
|
|
id: number;
|
|
memory_session_id: string;
|
|
project: string;
|
|
merged_into_project: string | null;
|
|
request: string | null;
|
|
investigated: string | null;
|
|
learned: string | null;
|
|
completed: string | null;
|
|
next_steps: string | null;
|
|
notes: string | null;
|
|
prompt_number: number;
|
|
discovery_tokens: number; // ROI metrics
|
|
created_at: string;
|
|
created_at_epoch: number;
|
|
}
|
|
|
|
interface StoredUserPrompt {
|
|
id: number;
|
|
content_session_id: string;
|
|
prompt_number: number;
|
|
prompt_text: string;
|
|
created_at: string;
|
|
created_at_epoch: number;
|
|
memory_session_id: string;
|
|
project: string;
|
|
}
|
|
|
|
export class ChromaSync {
|
|
private project: string;
|
|
private collectionName: string;
|
|
private collectionCreated = false;
|
|
private readonly BATCH_SIZE = 100;
|
|
|
|
constructor(project: string) {
|
|
this.project = project;
|
|
// Chroma collection names only allow [a-zA-Z0-9._-], 3-512 chars,
|
|
// must start/end with [a-zA-Z0-9]
|
|
const sanitized = project
|
|
.replace(/[^a-zA-Z0-9._-]/g, '_')
|
|
.replace(/[^a-zA-Z0-9]+$/, ''); // strip trailing non-alphanumeric
|
|
this.collectionName = `cm__${sanitized || 'unknown'}`;
|
|
}
|
|
|
|
/**
|
|
* Ensure collection exists in Chroma via MCP.
|
|
* chroma_create_collection is idempotent - safe to call multiple times.
|
|
* Uses collectionCreated flag to avoid redundant calls within a session.
|
|
*/
|
|
private async ensureCollectionExists(): Promise<void> {
|
|
if (this.collectionCreated) {
|
|
return;
|
|
}
|
|
|
|
const chromaMcp = ChromaMcpManager.getInstance();
|
|
try {
|
|
await chromaMcp.callTool('chroma_create_collection', {
|
|
collection_name: this.collectionName
|
|
});
|
|
} catch (error) {
|
|
const message = error instanceof Error ? error.message : String(error);
|
|
if (!message.includes('already exists')) {
|
|
throw error;
|
|
}
|
|
// Collection already exists - this is the expected path after first creation
|
|
}
|
|
|
|
this.collectionCreated = true;
|
|
|
|
logger.debug('CHROMA_SYNC', 'Collection ready', {
|
|
collection: this.collectionName
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Format observation into Chroma documents (granular approach)
|
|
* Each semantic field becomes a separate vector document
|
|
*/
|
|
private formatObservationDocs(obs: StoredObservation): ChromaDocument[] {
|
|
const documents: ChromaDocument[] = [];
|
|
|
|
// Parse JSON fields
|
|
const facts = obs.facts ? JSON.parse(obs.facts) : [];
|
|
const concepts = obs.concepts ? JSON.parse(obs.concepts) : [];
|
|
const files_read = parseFileList(obs.files_read);
|
|
const files_modified = parseFileList(obs.files_modified);
|
|
|
|
const baseMetadata: Record<string, string | number | null> = {
|
|
sqlite_id: obs.id,
|
|
doc_type: 'observation',
|
|
memory_session_id: obs.memory_session_id,
|
|
project: obs.project,
|
|
merged_into_project: obs.merged_into_project ?? null,
|
|
created_at_epoch: obs.created_at_epoch,
|
|
type: obs.type || 'discovery',
|
|
title: obs.title || 'Untitled'
|
|
};
|
|
|
|
// Add optional metadata fields
|
|
if (obs.subtitle) {
|
|
baseMetadata.subtitle = obs.subtitle;
|
|
}
|
|
if (concepts.length > 0) {
|
|
baseMetadata.concepts = concepts.join(',');
|
|
}
|
|
if (files_read.length > 0) {
|
|
baseMetadata.files_read = files_read.join(',');
|
|
}
|
|
if (files_modified.length > 0) {
|
|
baseMetadata.files_modified = files_modified.join(',');
|
|
}
|
|
|
|
// Narrative as separate document
|
|
if (obs.narrative) {
|
|
documents.push({
|
|
id: `obs_${obs.id}_narrative`,
|
|
document: obs.narrative,
|
|
metadata: { ...baseMetadata, field_type: 'narrative' }
|
|
});
|
|
}
|
|
|
|
// Text as separate document (legacy field)
|
|
if (obs.text) {
|
|
documents.push({
|
|
id: `obs_${obs.id}_text`,
|
|
document: obs.text,
|
|
metadata: { ...baseMetadata, field_type: 'text' }
|
|
});
|
|
}
|
|
|
|
// Each fact as separate document
|
|
facts.forEach((fact: string, index: number) => {
|
|
documents.push({
|
|
id: `obs_${obs.id}_fact_${index}`,
|
|
document: fact,
|
|
metadata: { ...baseMetadata, field_type: 'fact', fact_index: index }
|
|
});
|
|
});
|
|
|
|
return documents;
|
|
}
|
|
|
|
/**
|
|
* Format summary into Chroma documents (granular approach)
|
|
* Each summary field becomes a separate vector document
|
|
*/
|
|
private formatSummaryDocs(summary: StoredSummary): ChromaDocument[] {
|
|
const documents: ChromaDocument[] = [];
|
|
|
|
const baseMetadata: Record<string, string | number | null> = {
|
|
sqlite_id: summary.id,
|
|
doc_type: 'session_summary',
|
|
memory_session_id: summary.memory_session_id,
|
|
project: summary.project,
|
|
merged_into_project: summary.merged_into_project ?? null,
|
|
created_at_epoch: summary.created_at_epoch,
|
|
prompt_number: summary.prompt_number || 0
|
|
};
|
|
|
|
// Each field becomes a separate document
|
|
if (summary.request) {
|
|
documents.push({
|
|
id: `summary_${summary.id}_request`,
|
|
document: summary.request,
|
|
metadata: { ...baseMetadata, field_type: 'request' }
|
|
});
|
|
}
|
|
|
|
if (summary.investigated) {
|
|
documents.push({
|
|
id: `summary_${summary.id}_investigated`,
|
|
document: summary.investigated,
|
|
metadata: { ...baseMetadata, field_type: 'investigated' }
|
|
});
|
|
}
|
|
|
|
if (summary.learned) {
|
|
documents.push({
|
|
id: `summary_${summary.id}_learned`,
|
|
document: summary.learned,
|
|
metadata: { ...baseMetadata, field_type: 'learned' }
|
|
});
|
|
}
|
|
|
|
if (summary.completed) {
|
|
documents.push({
|
|
id: `summary_${summary.id}_completed`,
|
|
document: summary.completed,
|
|
metadata: { ...baseMetadata, field_type: 'completed' }
|
|
});
|
|
}
|
|
|
|
if (summary.next_steps) {
|
|
documents.push({
|
|
id: `summary_${summary.id}_next_steps`,
|
|
document: summary.next_steps,
|
|
metadata: { ...baseMetadata, field_type: 'next_steps' }
|
|
});
|
|
}
|
|
|
|
if (summary.notes) {
|
|
documents.push({
|
|
id: `summary_${summary.id}_notes`,
|
|
document: summary.notes,
|
|
metadata: { ...baseMetadata, field_type: 'notes' }
|
|
});
|
|
}
|
|
|
|
return documents;
|
|
}
|
|
|
|
/**
|
|
* Add documents to Chroma in batch via MCP
|
|
* Throws error if batch add fails
|
|
*/
|
|
private async addDocuments(documents: ChromaDocument[]): Promise<void> {
|
|
if (documents.length === 0) {
|
|
return;
|
|
}
|
|
|
|
await this.ensureCollectionExists();
|
|
|
|
const chromaMcp = ChromaMcpManager.getInstance();
|
|
|
|
// Add in batches
|
|
for (let i = 0; i < documents.length; i += this.BATCH_SIZE) {
|
|
const batch = documents.slice(i, i + this.BATCH_SIZE);
|
|
|
|
// Sanitize metadata: filter out null, undefined, and empty string values
|
|
// that chroma-mcp may reject (e.g., null subtitle from raw SQLite rows)
|
|
const cleanMetadatas = batch.map(d =>
|
|
Object.fromEntries(
|
|
Object.entries(d.metadata).filter(([_, v]) => v !== null && v !== undefined && v !== '')
|
|
)
|
|
);
|
|
|
|
try {
|
|
await chromaMcp.callTool('chroma_add_documents', {
|
|
collection_name: this.collectionName,
|
|
ids: batch.map(d => d.id),
|
|
documents: batch.map(d => d.document),
|
|
metadatas: cleanMetadatas
|
|
});
|
|
} catch (error) {
|
|
const errMsg = error instanceof Error ? error.message : String(error);
|
|
// APPROVED OVERRIDE: Duplicate IDs from partial write before timeout/crash.
|
|
// chroma_update_documents only updates *existing* IDs — it silently ignores
|
|
// missing ones. So we delete-then-add to guarantee all IDs are written.
|
|
if (errMsg.includes('already exist')) {
|
|
try {
|
|
await chromaMcp.callTool('chroma_delete_documents', {
|
|
collection_name: this.collectionName,
|
|
ids: batch.map(d => d.id)
|
|
});
|
|
await chromaMcp.callTool('chroma_add_documents', {
|
|
collection_name: this.collectionName,
|
|
ids: batch.map(d => d.id),
|
|
documents: batch.map(d => d.document),
|
|
metadatas: cleanMetadatas
|
|
});
|
|
logger.info('CHROMA_SYNC', 'Batch reconciled via delete+add after duplicate conflict', {
|
|
collection: this.collectionName,
|
|
batchStart: i,
|
|
batchSize: batch.length
|
|
});
|
|
} catch (reconcileError) {
|
|
logger.error('CHROMA_SYNC', 'Batch reconcile (delete+add) failed', {
|
|
collection: this.collectionName,
|
|
batchStart: i,
|
|
batchSize: batch.length
|
|
}, reconcileError as Error);
|
|
}
|
|
} else {
|
|
logger.error('CHROMA_SYNC', 'Batch add failed, continuing with remaining batches', {
|
|
collection: this.collectionName,
|
|
batchStart: i,
|
|
batchSize: batch.length
|
|
}, error as Error);
|
|
}
|
|
}
|
|
}
|
|
|
|
logger.debug('CHROMA_SYNC', 'Documents added', {
|
|
collection: this.collectionName,
|
|
count: documents.length
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Sync a single observation to Chroma
|
|
* Blocks until sync completes, throws on error
|
|
*/
|
|
async syncObservation(
|
|
observationId: number,
|
|
memorySessionId: string,
|
|
project: string,
|
|
obs: ParsedObservation,
|
|
promptNumber: number,
|
|
createdAtEpoch: number,
|
|
discoveryTokens: number = 0
|
|
): Promise<void> {
|
|
// Convert ParsedObservation to StoredObservation format
|
|
const stored: StoredObservation = {
|
|
id: observationId,
|
|
memory_session_id: memorySessionId,
|
|
project: project,
|
|
merged_into_project: null,
|
|
text: null, // Legacy field, not used
|
|
type: obs.type,
|
|
title: obs.title,
|
|
subtitle: obs.subtitle,
|
|
facts: JSON.stringify(obs.facts),
|
|
narrative: obs.narrative,
|
|
concepts: JSON.stringify(obs.concepts),
|
|
files_read: JSON.stringify(obs.files_read),
|
|
files_modified: JSON.stringify(obs.files_modified),
|
|
prompt_number: promptNumber,
|
|
discovery_tokens: discoveryTokens,
|
|
created_at: new Date(createdAtEpoch * 1000).toISOString(),
|
|
created_at_epoch: createdAtEpoch
|
|
};
|
|
|
|
const documents = this.formatObservationDocs(stored);
|
|
|
|
logger.info('CHROMA_SYNC', 'Syncing observation', {
|
|
observationId,
|
|
documentCount: documents.length,
|
|
project
|
|
});
|
|
|
|
await this.addDocuments(documents);
|
|
}
|
|
|
|
/**
|
|
* Sync a single summary to Chroma
|
|
* Blocks until sync completes, throws on error
|
|
*/
|
|
async syncSummary(
|
|
summaryId: number,
|
|
memorySessionId: string,
|
|
project: string,
|
|
summary: ParsedSummary,
|
|
promptNumber: number,
|
|
createdAtEpoch: number,
|
|
discoveryTokens: number = 0
|
|
): Promise<void> {
|
|
// Convert ParsedSummary to StoredSummary format
|
|
const stored: StoredSummary = {
|
|
id: summaryId,
|
|
memory_session_id: memorySessionId,
|
|
project: project,
|
|
merged_into_project: null,
|
|
request: summary.request,
|
|
investigated: summary.investigated,
|
|
learned: summary.learned,
|
|
completed: summary.completed,
|
|
next_steps: summary.next_steps,
|
|
notes: summary.notes,
|
|
prompt_number: promptNumber,
|
|
discovery_tokens: discoveryTokens,
|
|
created_at: new Date(createdAtEpoch * 1000).toISOString(),
|
|
created_at_epoch: createdAtEpoch
|
|
};
|
|
|
|
const documents = this.formatSummaryDocs(stored);
|
|
|
|
logger.info('CHROMA_SYNC', 'Syncing summary', {
|
|
summaryId,
|
|
documentCount: documents.length,
|
|
project
|
|
});
|
|
|
|
await this.addDocuments(documents);
|
|
}
|
|
|
|
/**
|
|
* Format user prompt into Chroma document
|
|
* Each prompt becomes a single document (unlike observations/summaries which split by field)
|
|
*/
|
|
private formatUserPromptDoc(prompt: StoredUserPrompt): ChromaDocument {
|
|
return {
|
|
id: `prompt_${prompt.id}`,
|
|
document: prompt.prompt_text,
|
|
metadata: {
|
|
sqlite_id: prompt.id,
|
|
doc_type: 'user_prompt',
|
|
memory_session_id: prompt.memory_session_id,
|
|
project: prompt.project,
|
|
created_at_epoch: prompt.created_at_epoch,
|
|
prompt_number: prompt.prompt_number
|
|
}
|
|
};
|
|
}
|
|
|
|
/**
|
|
* Sync a single user prompt to Chroma
|
|
* Blocks until sync completes, throws on error
|
|
*/
|
|
async syncUserPrompt(
|
|
promptId: number,
|
|
memorySessionId: string,
|
|
project: string,
|
|
promptText: string,
|
|
promptNumber: number,
|
|
createdAtEpoch: number
|
|
): Promise<void> {
|
|
// Create StoredUserPrompt format
|
|
const stored: StoredUserPrompt = {
|
|
id: promptId,
|
|
content_session_id: '', // Not needed for Chroma sync
|
|
prompt_number: promptNumber,
|
|
prompt_text: promptText,
|
|
created_at: new Date(createdAtEpoch * 1000).toISOString(),
|
|
created_at_epoch: createdAtEpoch,
|
|
memory_session_id: memorySessionId,
|
|
project: project
|
|
};
|
|
|
|
const document = this.formatUserPromptDoc(stored);
|
|
|
|
logger.info('CHROMA_SYNC', 'Syncing user prompt', {
|
|
promptId,
|
|
project
|
|
});
|
|
|
|
await this.addDocuments([document]);
|
|
}
|
|
|
|
/**
|
|
* Fetch all existing document IDs from Chroma collection via MCP
|
|
* Returns Sets of SQLite IDs for observations, summaries, and prompts
|
|
*/
|
|
private async getExistingChromaIds(projectOverride?: string): Promise<{
|
|
observations: Set<number>;
|
|
summaries: Set<number>;
|
|
prompts: Set<number>;
|
|
}> {
|
|
const targetProject = projectOverride ?? this.project;
|
|
await this.ensureCollectionExists();
|
|
|
|
const chromaMcp = ChromaMcpManager.getInstance();
|
|
|
|
const observationIds = new Set<number>();
|
|
const summaryIds = new Set<number>();
|
|
const promptIds = new Set<number>();
|
|
|
|
let offset = 0;
|
|
const limit = 1000; // Large batches, metadata only = fast
|
|
|
|
logger.info('CHROMA_SYNC', 'Fetching existing Chroma document IDs...', { project: targetProject });
|
|
|
|
while (true) {
|
|
const result = await chromaMcp.callTool('chroma_get_documents', {
|
|
collection_name: this.collectionName,
|
|
limit: limit,
|
|
offset: offset,
|
|
where: { project: targetProject },
|
|
include: ['metadatas']
|
|
}) as any;
|
|
|
|
// chroma_get_documents returns flat arrays: { ids, metadatas, documents }
|
|
const metadatas = result?.metadatas || [];
|
|
|
|
if (metadatas.length === 0) {
|
|
break; // No more documents
|
|
}
|
|
|
|
// Extract SQLite IDs from metadata
|
|
for (const meta of metadatas) {
|
|
if (meta && meta.sqlite_id) {
|
|
const sqliteId = meta.sqlite_id as number;
|
|
if (meta.doc_type === 'observation') {
|
|
observationIds.add(sqliteId);
|
|
} else if (meta.doc_type === 'session_summary') {
|
|
summaryIds.add(sqliteId);
|
|
} else if (meta.doc_type === 'user_prompt') {
|
|
promptIds.add(sqliteId);
|
|
}
|
|
}
|
|
}
|
|
|
|
offset += limit;
|
|
|
|
logger.debug('CHROMA_SYNC', 'Fetched batch of existing IDs', {
|
|
project: targetProject,
|
|
offset,
|
|
batchSize: metadatas.length
|
|
});
|
|
}
|
|
|
|
logger.info('CHROMA_SYNC', 'Existing IDs fetched', {
|
|
project: targetProject,
|
|
observations: observationIds.size,
|
|
summaries: summaryIds.size,
|
|
prompts: promptIds.size
|
|
});
|
|
|
|
return { observations: observationIds, summaries: summaryIds, prompts: promptIds };
|
|
}
|
|
|
|
/**
|
|
* Backfill: Sync all observations missing from Chroma
|
|
* Reads from SQLite and syncs in batches
|
|
* @param projectOverride - If provided, backfill this project instead of this.project.
|
|
* Used by backfillAllProjects() to iterate projects without mutating instance state.
|
|
* Throws error if backfill fails
|
|
*/
|
|
async ensureBackfilled(projectOverride?: string): Promise<void> {
|
|
const backfillProject = projectOverride ?? this.project;
|
|
logger.info('CHROMA_SYNC', 'Starting smart backfill', { project: backfillProject });
|
|
|
|
await this.ensureCollectionExists();
|
|
|
|
// Fetch existing IDs from Chroma (fast, metadata only)
|
|
const existing = await this.getExistingChromaIds(backfillProject);
|
|
|
|
const db = new SessionStore();
|
|
|
|
try {
|
|
// Build exclusion list for observations
|
|
// Filter to validated positive integers before interpolating into SQL
|
|
const existingObsIds = Array.from(existing.observations).filter(id => Number.isInteger(id) && id > 0);
|
|
const obsExclusionClause = existingObsIds.length > 0
|
|
? `AND id NOT IN (${existingObsIds.join(',')})`
|
|
: '';
|
|
|
|
// Get only observations missing from Chroma
|
|
const observations = db.db.prepare(`
|
|
SELECT * FROM observations
|
|
WHERE project = ? ${obsExclusionClause}
|
|
ORDER BY id ASC
|
|
`).all(backfillProject) as StoredObservation[];
|
|
|
|
const totalObsCount = db.db.prepare(`
|
|
SELECT COUNT(*) as count FROM observations WHERE project = ?
|
|
`).get(backfillProject) as { count: number };
|
|
|
|
logger.info('CHROMA_SYNC', 'Backfilling observations', {
|
|
project: backfillProject,
|
|
missing: observations.length,
|
|
existing: existing.observations.size,
|
|
total: totalObsCount.count
|
|
});
|
|
|
|
// Format all observation documents
|
|
const allDocs: ChromaDocument[] = [];
|
|
for (const obs of observations) {
|
|
allDocs.push(...this.formatObservationDocs(obs));
|
|
}
|
|
|
|
// Sync in batches
|
|
for (let i = 0; i < allDocs.length; i += this.BATCH_SIZE) {
|
|
const batch = allDocs.slice(i, i + this.BATCH_SIZE);
|
|
await this.addDocuments(batch);
|
|
|
|
logger.debug('CHROMA_SYNC', 'Backfill progress', {
|
|
project: backfillProject,
|
|
progress: `${Math.min(i + this.BATCH_SIZE, allDocs.length)}/${allDocs.length}`
|
|
});
|
|
}
|
|
|
|
// Build exclusion list for summaries
|
|
const existingSummaryIds = Array.from(existing.summaries).filter(id => Number.isInteger(id) && id > 0);
|
|
const summaryExclusionClause = existingSummaryIds.length > 0
|
|
? `AND id NOT IN (${existingSummaryIds.join(',')})`
|
|
: '';
|
|
|
|
// Get only summaries missing from Chroma
|
|
const summaries = db.db.prepare(`
|
|
SELECT * FROM session_summaries
|
|
WHERE project = ? ${summaryExclusionClause}
|
|
ORDER BY id ASC
|
|
`).all(backfillProject) as StoredSummary[];
|
|
|
|
const totalSummaryCount = db.db.prepare(`
|
|
SELECT COUNT(*) as count FROM session_summaries WHERE project = ?
|
|
`).get(backfillProject) as { count: number };
|
|
|
|
logger.info('CHROMA_SYNC', 'Backfilling summaries', {
|
|
project: backfillProject,
|
|
missing: summaries.length,
|
|
existing: existing.summaries.size,
|
|
total: totalSummaryCount.count
|
|
});
|
|
|
|
// Format all summary documents
|
|
const summaryDocs: ChromaDocument[] = [];
|
|
for (const summary of summaries) {
|
|
summaryDocs.push(...this.formatSummaryDocs(summary));
|
|
}
|
|
|
|
// Sync in batches
|
|
for (let i = 0; i < summaryDocs.length; i += this.BATCH_SIZE) {
|
|
const batch = summaryDocs.slice(i, i + this.BATCH_SIZE);
|
|
await this.addDocuments(batch);
|
|
|
|
logger.debug('CHROMA_SYNC', 'Backfill progress', {
|
|
project: backfillProject,
|
|
progress: `${Math.min(i + this.BATCH_SIZE, summaryDocs.length)}/${summaryDocs.length}`
|
|
});
|
|
}
|
|
|
|
// Build exclusion list for prompts
|
|
const existingPromptIds = Array.from(existing.prompts).filter(id => Number.isInteger(id) && id > 0);
|
|
const promptExclusionClause = existingPromptIds.length > 0
|
|
? `AND up.id NOT IN (${existingPromptIds.join(',')})`
|
|
: '';
|
|
|
|
// Get only user prompts missing from Chroma
|
|
const prompts = db.db.prepare(`
|
|
SELECT
|
|
up.*,
|
|
s.project,
|
|
s.memory_session_id
|
|
FROM user_prompts up
|
|
JOIN sdk_sessions s ON up.content_session_id = s.content_session_id
|
|
WHERE s.project = ? ${promptExclusionClause}
|
|
ORDER BY up.id ASC
|
|
`).all(backfillProject) as StoredUserPrompt[];
|
|
|
|
const totalPromptCount = db.db.prepare(`
|
|
SELECT COUNT(*) as count
|
|
FROM user_prompts up
|
|
JOIN sdk_sessions s ON up.content_session_id = s.content_session_id
|
|
WHERE s.project = ?
|
|
`).get(backfillProject) as { count: number };
|
|
|
|
logger.info('CHROMA_SYNC', 'Backfilling user prompts', {
|
|
project: backfillProject,
|
|
missing: prompts.length,
|
|
existing: existing.prompts.size,
|
|
total: totalPromptCount.count
|
|
});
|
|
|
|
// Format all prompt documents
|
|
const promptDocs: ChromaDocument[] = [];
|
|
for (const prompt of prompts) {
|
|
promptDocs.push(this.formatUserPromptDoc(prompt));
|
|
}
|
|
|
|
// Sync in batches
|
|
for (let i = 0; i < promptDocs.length; i += this.BATCH_SIZE) {
|
|
const batch = promptDocs.slice(i, i + this.BATCH_SIZE);
|
|
await this.addDocuments(batch);
|
|
|
|
logger.debug('CHROMA_SYNC', 'Backfill progress', {
|
|
project: backfillProject,
|
|
progress: `${Math.min(i + this.BATCH_SIZE, promptDocs.length)}/${promptDocs.length}`
|
|
});
|
|
}
|
|
|
|
logger.info('CHROMA_SYNC', 'Smart backfill complete', {
|
|
project: backfillProject,
|
|
synced: {
|
|
observationDocs: allDocs.length,
|
|
summaryDocs: summaryDocs.length,
|
|
promptDocs: promptDocs.length
|
|
},
|
|
skipped: {
|
|
observations: existing.observations.size,
|
|
summaries: existing.summaries.size,
|
|
prompts: existing.prompts.size
|
|
}
|
|
});
|
|
|
|
} catch (error) {
|
|
logger.error('CHROMA_SYNC', 'Backfill failed', { project: backfillProject }, error as Error);
|
|
throw new Error(`Backfill failed: ${error instanceof Error ? error.message : String(error)}`);
|
|
} finally {
|
|
db.close();
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Query Chroma collection for semantic search via MCP
|
|
* Used by SearchManager for vector-based search
|
|
*/
|
|
async queryChroma(
|
|
query: string,
|
|
limit: number,
|
|
whereFilter?: Record<string, any>
|
|
): Promise<{ ids: number[]; distances: number[]; metadatas: any[] }> {
|
|
await this.ensureCollectionExists();
|
|
|
|
try {
|
|
const chromaMcp = ChromaMcpManager.getInstance();
|
|
const results = await chromaMcp.callTool('chroma_query_documents', {
|
|
collection_name: this.collectionName,
|
|
query_texts: [query],
|
|
n_results: limit,
|
|
...(whereFilter && { where: whereFilter }),
|
|
include: ['documents', 'metadatas', 'distances']
|
|
}) as any;
|
|
|
|
// chroma_query_documents returns nested arrays (one per query text)
|
|
// We always pass a single query text, so we access [0]
|
|
const ids: number[] = [];
|
|
const seen = new Set<number>();
|
|
const docIds = results?.ids?.[0] || [];
|
|
const rawMetadatas = results?.metadatas?.[0] || [];
|
|
const rawDistances = results?.distances?.[0] || [];
|
|
|
|
// Build deduplicated arrays that stay index-aligned:
|
|
// Multiple Chroma docs map to the same SQLite ID (one per field).
|
|
// Keep the first (best-ranked) distance and metadata per SQLite ID.
|
|
const metadatas: any[] = [];
|
|
const distances: number[] = [];
|
|
|
|
for (let i = 0; i < docIds.length; i++) {
|
|
const docId = docIds[i];
|
|
// Extract sqlite_id from document ID (supports three formats):
|
|
// - obs_{id}_narrative, obs_{id}_fact_0, etc (observations)
|
|
// - summary_{id}_request, summary_{id}_learned, etc (session summaries)
|
|
// - prompt_{id} (user prompts)
|
|
const obsMatch = docId.match(/obs_(\d+)_/);
|
|
const summaryMatch = docId.match(/summary_(\d+)_/);
|
|
const promptMatch = docId.match(/prompt_(\d+)/);
|
|
|
|
let sqliteId: number | null = null;
|
|
if (obsMatch) {
|
|
sqliteId = parseInt(obsMatch[1], 10);
|
|
} else if (summaryMatch) {
|
|
sqliteId = parseInt(summaryMatch[1], 10);
|
|
} else if (promptMatch) {
|
|
sqliteId = parseInt(promptMatch[1], 10);
|
|
}
|
|
|
|
if (sqliteId !== null && !seen.has(sqliteId)) {
|
|
seen.add(sqliteId);
|
|
ids.push(sqliteId);
|
|
metadatas.push(rawMetadatas[i] ?? null);
|
|
distances.push(rawDistances[i] ?? 0);
|
|
}
|
|
}
|
|
|
|
return { ids, distances, metadatas };
|
|
} catch (error) {
|
|
const errorMessage = error instanceof Error ? error.message : String(error);
|
|
|
|
// Check for connection errors
|
|
const isConnectionError =
|
|
errorMessage.includes('ECONNREFUSED') ||
|
|
errorMessage.includes('ENOTFOUND') ||
|
|
errorMessage.includes('fetch failed') ||
|
|
errorMessage.includes('subprocess closed') ||
|
|
errorMessage.includes('timed out');
|
|
|
|
if (isConnectionError) {
|
|
// Reset collection state so next call attempts reconnect
|
|
this.collectionCreated = false;
|
|
logger.error('CHROMA_SYNC', 'Connection lost during query',
|
|
{ project: this.project, query }, error as Error);
|
|
throw new Error(`Chroma query failed - connection lost: ${errorMessage}`);
|
|
}
|
|
|
|
logger.error('CHROMA_SYNC', 'Query failed', { project: this.project, query }, error as Error);
|
|
throw error;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Backfill all projects that have observations in SQLite but may be missing from Chroma.
|
|
* Uses a single shared ChromaSync('claude-mem') instance and Chroma connection.
|
|
* Per-project scoping is passed as a parameter to ensureBackfilled(), avoiding
|
|
* instance state mutation. All documents land in the cm__claude-mem collection
|
|
* with project scoped via metadata, matching how DatabaseManager and SearchManager operate.
|
|
* Designed to be called fire-and-forget on worker startup.
|
|
*/
|
|
static async backfillAllProjects(): Promise<void> {
|
|
const db = new SessionStore();
|
|
const sync = new ChromaSync('claude-mem');
|
|
try {
|
|
const projects = db.db.prepare(
|
|
'SELECT DISTINCT project FROM observations WHERE project IS NOT NULL AND project != ?'
|
|
).all('') as { project: string }[];
|
|
|
|
logger.info('CHROMA_SYNC', `Backfill check for ${projects.length} projects`);
|
|
|
|
for (const { project } of projects) {
|
|
try {
|
|
await sync.ensureBackfilled(project);
|
|
} catch (error) {
|
|
logger.error('CHROMA_SYNC', `Backfill failed for project: ${project}`, {}, error as Error);
|
|
// Continue to next project — don't let one failure stop others
|
|
}
|
|
}
|
|
} finally {
|
|
await sync.close();
|
|
db.close();
|
|
}
|
|
}
|
|
|
|
/**
|
|
* Stamp `merged_into_project` on every Chroma document whose metadata
|
|
* `sqlite_id` is in the provided set. Used by the worktree adoption engine
|
|
* to keep Chroma's metadata in lockstep with SQLite after a parent branch
|
|
* absorbs a worktree branch via merge.
|
|
*
|
|
* Batched: fetches docs by `sqlite_id IN sqliteIds`, rewrites metadata with
|
|
* the new field, and calls `chroma_update_documents` once per page of up to
|
|
* BATCH_SIZE ids. Idempotent — re-running with the same value is a no-op
|
|
* because the write doesn't depend on the prior value.
|
|
*/
|
|
async updateMergedIntoProject(
|
|
sqliteIds: number[],
|
|
mergedIntoProject: string
|
|
): Promise<void> {
|
|
if (sqliteIds.length === 0) return;
|
|
|
|
await this.ensureCollectionExists();
|
|
const chromaMcp = ChromaMcpManager.getInstance();
|
|
|
|
let totalPatched = 0;
|
|
|
|
// Chunk the sqlite_id set to keep each Chroma call bounded.
|
|
for (let i = 0; i < sqliteIds.length; i += this.BATCH_SIZE) {
|
|
const idBatch = sqliteIds.slice(i, i + this.BATCH_SIZE);
|
|
|
|
const existing = await chromaMcp.callTool('chroma_get_documents', {
|
|
collection_name: this.collectionName,
|
|
where: { sqlite_id: { $in: idBatch } },
|
|
include: ['metadatas']
|
|
}) as { ids?: string[]; metadatas?: Array<Record<string, any> | null> };
|
|
|
|
const docIds: string[] = existing?.ids ?? [];
|
|
if (docIds.length === 0) continue;
|
|
|
|
const metadatas = (existing?.metadatas ?? []).map(m => {
|
|
// Merge old metadata with the new field, then filter out null/undefined/''
|
|
// to match the sanitization other callTool sites apply (chroma-mcp
|
|
// rejects null values in metadata).
|
|
const merged: Record<string, any> = {
|
|
...(m ?? {}),
|
|
merged_into_project: mergedIntoProject
|
|
};
|
|
return Object.fromEntries(
|
|
Object.entries(merged).filter(
|
|
([, v]) => v !== null && v !== undefined && v !== ''
|
|
)
|
|
);
|
|
});
|
|
|
|
await chromaMcp.callTool('chroma_update_documents', {
|
|
collection_name: this.collectionName,
|
|
ids: docIds,
|
|
metadatas
|
|
});
|
|
totalPatched += docIds.length;
|
|
}
|
|
|
|
logger.info('CHROMA_SYNC', 'merged_into_project metadata patched', {
|
|
collection: this.collectionName,
|
|
mergedIntoProject,
|
|
sqliteIdCount: sqliteIds.length,
|
|
chromaDocsPatched: totalPatched
|
|
});
|
|
}
|
|
|
|
/**
|
|
* Close the ChromaSync instance
|
|
* ChromaMcpManager is a singleton and manages its own lifecycle
|
|
* We don't close it here - it's closed during graceful shutdown
|
|
*/
|
|
async close(): Promise<void> {
|
|
// ChromaMcpManager is a singleton and manages its own lifecycle
|
|
// We don't close it here - it's closed during graceful shutdown
|
|
logger.info('CHROMA_SYNC', 'ChromaSync closed', { project: this.project });
|
|
}
|
|
}
|