| name | automem-search |
| description | Search, filter, and retrieve Claude/Codex history indexed by the automem CLI. Use when the user wants to index history, run lexical/semantic/hybrid search, fetch full transcripts, or produce LLM-friendly JSON output for RAG. |
Automem Search
Use this skill to index local history and retrieve results in a structured, LLM-friendly way.
Indexing
- Build or update the index (incremental):
./target/debug/automem index
- Full rebuild (clears index):
./target/debug/automem reindex
- Embeddings are on by default.
- Disable embeddings:
./target/debug/automem index --no-embeddings
- Backfill embeddings only:
./target/debug/automem embed
- Common flags:
--source <path>for Claude logs--include-agentsto include agent transcripts--codex/--no-codexto include or skip Codex logs--root <path>to change data root (default:~/.automem)
Search (LLM default JSON)
Run a search; output is JSON lines by default.
./target/debug/automem search "query" --limit 20
Each JSON line includes:
doc_id,ts(ISO),session_id,project,role,source_pathtext(full record text)snippet(trimmed single-line summary)matches(offsets + before/after context)score(ranked score)
Mode decision table
| Need | Command |
|---|---|
| Exact terms | search "exact term" |
| Fuzzy concepts | search "concept" --semantic |
| Mixed | search "term concept" --hybrid |
Filters
--project <name>--role <user|assistant|tool_use|tool_result>--tool <tool_name>--session <session_id>(search inside a transcript)--source claude|codex--since <iso|unix>/--until <iso|unix>--limit <n>--min-score <float>
Grouping / dedupe
--top-n-per-session <n>(top n per session)--unique-session(same as top‑k per session = 1)--sort score|ts(default score)
Output shape
- JSONL default (one JSON per line)
--json-arrayfor a single JSON array--fields score,ts,doc_id,session_id,snippetto reduce output-v/--verbosefor human output
Narrow first (fastest reducers)
- Global search with
--limit - Reduce with
--projectand--since/--until - Optionally
--top-n-per-sessionor--unique-session ./target/debug/automem session <id>for full context
Practical narrowing tips
- Start with exact terms (quoted) before hybrid if results are noisy.
- Use
--unique-sessionto collapse PR‑link spam fast. - Use
--min-scoreto prune low-signal hits. - Use
--sort tswhen you want a timeline view. - Use
--role assistantfor narrative outcomes;--role tool_resultfor command errors. - For a specific session, prefer
search "<term>" --session <id> --sort ts --limit 50to jump to outcomes.
Config
Create ~/.automem/config.toml (or <root>/config.toml if you use --root):
embeddings = true
auto_index_on_search = true
auto_index_on_search runs an incremental index update before each search.
Semantic and Hybrid
- Semantic:
--semantic - Hybrid (BM25 + vectors, RRF):
--hybrid - Recency tuning:
--recency-weight <float>--recency-half-life-days <float>
Fetch Full Context
- One record:
./target/debug/automem show <doc_id>
- Full transcript:
./target/debug/automem session <session_id>
Both commands return JSON by default.
Human Output
Use -v/--verbose for human-readable output:
./target/debug/automem search "query" -v./target/debug/automem show <doc_id> -v./target/debug/automem session <session_id> -v
Recommended LLM Flow
./target/debug/automem search "query" --limit 20- Pick hits using
matchesorsnippet ./target/debug/automem show <doc_id>or./target/debug/automem session <session_id>- Refine with
--session,--role, or time filters