How it works
Three pillars. One MCP server.
CodeSift sits between your AI agent and your codebase. It parses code into an AST index, ranks results with BM25F, and bridges to language servers — delivering 61% fewer tokens while giving agents capabilities grep cannot provide.
Without CodeSift
rg "auth" — raw text matches rg "middleware" — more raw matches With CodeSift
assemble_context("auth", level="L1") The three pillars
AST Index
Tree-sitter parses every file into an abstract syntax tree. Symbols are extracted, scored by centrality (how often imported), and indexed with BM25F for fast, ranked retrieval.
Semantic Search
Embeddings turn code into vectors. Questions like "how does authentication work?" find relevant code by meaning, not just matching keywords. Hybrid mode merges semantic + BM25 via Reciprocal Rank Fusion.
LSP Bridge
When a language server is available, CodeSift proxies type-aware operations: resolved definitions, hover types, cross-file rename. Lazy start, 5-minute idle kill, zero overhead when unused.
Two commands. Thirty seconds.
Install globally
$ npm install -g codesift-mcp One binary. No cloud. No signup. MIT licensed.
Add to your MCP config
{
"mcpServers": {
"codesift": {
"command": "codesift-mcp"
}
}
} Works with Claude Code, Cursor, Codex, and any MCP client.
Code normally
Your AI agent automatically discovers 64 MCP tools. It uses search_symbols instead of grep,
assemble_context instead of reading files, trace_route instead of guessing at endpoints.
You don't change how you work. The agent works better.
Measured savings
From real benchmarks across 188 agent sessions and 603 combo flow runs
| What the agent does | Native (grep/read) | CodeSift | Savings |
|---|---|---|---|
| Search for a symbol definition | ~57,000 tok | ~5,700 tok | -90% |
| Understand a feature ("how does auth work?") | ~93,000 tok | ~12,600 tok | -86% |
| Trace an HTTP route end-to-end | ~35,000 tok | ~61 tok | -99% |
| Scan for hardcoded secrets | ~1.6M tok | ~11,500 tok | -99% |
| Find unused exports (dead code) | 21 calls | 1 call | -82% |
| Real-world combo flows (13 sequences) | 4.58M tok | 1.86M tok | -61% |