Files
claude-mem/plugin/skills/mem-search/SKILL.md
T
Alex Newman 660c523ba4 fix: shorten MCP server name to prevent tool name length errors (#360)
* fix: shorten MCP server name to prevent tool name length errors (#358)

Root cause: Claude Code prefixes MCP tool names with
`mcp__plugin_{plugin-name}_{server-name}__` which was 43 chars
for `mcp__plugin_claude-mem_claude-mem-search__`. Combined with
`progressive_description` (22 chars) this exceeded the 64 char limit.

Changes:
- Shortened MCP server name from 'claude-mem-search' to 'mem-search'
  (saves 8 chars, new prefix is 35 chars)
- Renamed `progressive_description` tool to `help` (saves 18 chars)
- Updated SKILL.md to reference new `help` tool name
- Updated internal Server constructor name for consistency

All tool names now safely under 64 char limit:
- Longest is now `get_batch_observations` at 56 chars total
- `help` is only 39 chars total

Fixes #358

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>

* refactor: rename get_batch_observations to get_observations

The plural form naturally implies multiple items can be fetched,
following WordPress conventions. Simpler and clearer naming.

Also saves 6 additional characters for MCP tool name length.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>

* docs: update all references to renamed MCP tools

Updated documentation and code comments to reflect:
- progressive_description → help
- get_batch_observations → get_observations

Files updated:
- docs/public/usage/claude-desktop.mdx
- docs/public/architecture/worker-service.mdx
- src/services/worker/FormattingService.ts

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude Sonnet 4.5 <noreply@anthropic.com>

---------

Co-authored-by: Claude Sonnet 4.5 <noreply@anthropic.com>
2025-12-16 22:06:24 -05:00

5.1 KiB

name, description
name description
mem-search Search claude-mem's persistent cross-session memory database. Use when user asks "did we already solve this?", "how did we do X last time?", or needs work from previous sessions.

Memory Search

Search past work across all sessions. Simple workflow: search → get IDs → fetch details by ID.

When to Use

Use when users ask about PREVIOUS sessions (not current conversation):

  • "Did we already fix this?"
  • "How did we solve X last time?"
  • "What happened last week?"

The Workflow

ALWAYS follow this exact flow:

  1. Search - Get an index of results with IDs
  2. Timeline - Get context around top results to understand what was happening
  3. Review - Look at titles/dates/context, pick relevant IDs
  4. Fetch - Get full details ONLY for those IDs

Step 1: Search Everything

Use the search MCP tool:

Required parameters:

  • query - Search term
  • limit: 20 - You can request large indexes as necessary
  • project - Project name (required)

Example:

search(query="authentication", limit=20, project="my-project")

Returns:

| ID | Time | T | Title | Read | Work |
|----|------|---|-------|------|------|
| #11131 | 3:48 PM | 🟣 | Added JWT authentication | ~75 | 🛠️ 450 |
| #10942 | 2:15 PM | 🔴 | Fixed auth token expiration | ~50 | 🛠️ 200 |

Step 2: Get Timeline Context

You MUST understand "what was happening" around a result.

Use the timeline MCP tool:

Example with observation ID:

timeline(anchor=11131, depth_before=3, depth_after=3, project="my-project")

Example with query (finds anchor automatically):

timeline(query="authentication", depth_before=3, depth_after=3, project="my-project")

Returns exactly depth_before + 1 + depth_after items - observations, sessions, and prompts interleaved chronologically around the anchor.

When to use:

  • User asks "what was happening when..."
  • Need to understand sequence of events
  • Want broader context around a specific observation

Step 3: Pick IDs

Review the index results (and timeline if used). Identify which IDs are actually relevant. Discard the rest.

Step 4: Fetch by ID

For each relevant ID, fetch full details using MCP tools:

Fetch multiple observations (ALWAYS use for 2+ IDs):

get_observations(ids=[11131, 10942, 10855])

With ordering and limit:

get_observations(
  ids=[11131, 10942, 10855],
  orderBy="date_desc",
  limit=10,
  project="my-project"
)

Fetch single observation (only when fetching exactly 1):

get_observation(id=11131)

Fetch session:

get_session(id=2005)  # Just the number from S2005

Fetch prompt:

get_prompt(id=5421)

ID formats:

  • Observations: Just the number (11131)
  • Sessions: Just the number (2005) from "S2005"
  • Prompts: Just the number (5421)

Batch optimization:

  • ALWAYS use get_observations for 2+ observations
  • 10-100x more efficient than individual fetches
  • Single HTTP request vs N requests
  • Returns all results in one response
  • Supports ordering and filtering

Search Parameters

Basic:

  • query - What to search for (required)
  • limit - How many results (default 20)
  • project - Filter by project name (required)

Filters (optional):

  • type - Filter to "observations", "sessions", or "prompts"
  • dateStart - Start date (YYYY-MM-DD or epoch timestamp)
  • dateEnd - End date (YYYY-MM-DD or epoch timestamp)
  • obs_type - Filter observations by type (comma-separated): bugfix, feature, decision, discovery, change

Examples

Find recent bug fixes:

Use the search MCP tool with filters:

search(query="bug", type="observations", obs_type="bugfix", limit=20, project="my-project")

Find what happened last week:

Use date filters:

search(type="observations", dateStart="2025-11-11", limit=20, project="my-project")

Search everything:

Simple query search:

search(query="database migration", limit=20, project="my-project")

Get detailed instructions:

Use the help tool to load full instructions on-demand:

help(topic="workflow")  # Get 4-step workflow
help(topic="search_params")  # Get parameters reference
help(topic="examples")  # Get usage examples
help(topic="all")  # Get complete guide

Why This Workflow?

Token efficiency:

  • Search results: ~50-100 tokens per result (table index)
  • Full observation: ~500-1000 tokens each
  • 10x savings - only fetch full when you know it's relevant

Batch fetching:

  • Individual fetches: 10 HTTP requests, ~5-10s latency
  • Batch fetch: 1 HTTP request, ~0.5-1s latency
  • 10-100x faster for multi-observation queries

Clarity:

  • See everything first (table index)
  • Get timeline context around interesting results
  • Pick what matters based on context
  • Fetch details only for what you need (batch when possible)

Remember:

  • ALWAYS get timeline context to understand what was happening
  • ALWAYS use get_observations when fetching 2+ observations
  • The workflow is optimized: search → timeline → batch fetch = 10-100x faster