The leading, most token-efficient MCP server for GitHub source code exploration via tree-sitter AST parsing
-
Updated
Apr 23, 2026 - Python
The leading, most token-efficient MCP server for GitHub source code exploration via tree-sitter AST parsing
Universal AI context generator. Saves thousands of tokens per conversation in Claude Code, Cursor, Copilot, Codex, and more.
SDL-MCP gives coding agents the right code context, not your entire repo. It turns sprawling codebases into compact, high-signal context that saves tokens, speeds up workflows, and improves agent output.
MCP server that saves Claude Code tokens by delegating bounded tasks to local or cloud LLMs. Works with LM Studio, Ollama, vLLM, DeepSeek, Groq, Cerebras.
MCP server for Claude Code and Codex. One tool call replaces ~42 minutes of agent exploration
A reversible code minifier for AI. Save tokens by stripping code format in your prompt, then perfectly restore it in the responces.
Guardian Agent and Token Savings for Claude Code
Turn any OpenAPI spec into a native CLI binary. No MCP, no bloat, no runtime dependencies, ONLY CLI.
TSCG — Deterministic tool-schema compiler for LLM agents. 50-72% token savings, 50 tools in 2.4ms. Phi-4 recovers from 0% to 90% accuracy. 459 tests, zero dependencies, MIT.
Caveman output style for Claude Code: 40% fewer output tokens, always-on formatting
Run Claude Code or Codex 50–80% cheaper: Hermit is an MCP executor with Codex-first fallback routing (codex → z.ai → local) while your orchestrator stays in charge.
Make GitHub Copilot responses terse across VS Code Chat, Copilot CLI, cloud agent, and code review. One command. 40-75% fewer response tokens, no correctness hit.
Save Claude tokens by offloading tasks to free LLM APIs - 19x more workflows with the same token budget
lowfat - slim your command output. strips noise, saves tokens.
Local Ollama-backed MCP server for Claude Code. Query large files without burning context (~30x token savings)
CodeCrew — Senior engineering team for your AI-generated code. One command. Saves $20-40/day in tokens.
Local-first Model Context Protocol (MCP) memory layer for Codex CLI/Desktop, Claude Code, Gemini CLI, Qwen/DeepSeek/Ollama and agent workflows. SQLite + FTS5 compact context packs, token savings, read-only mode, no external memory server.
Nuke your token usage. Code indexing MCP server: 15 tools, 10 languages, O(1) retrieval, hybrid search, call graphs.
FigureOut is a Python package allows developers to easily integrate LLM into their application.
Add a description, image, and links to the token-savings topic page so that developers can more easily learn about it.
To associate your repository with the token-savings topic, visit your repo's landing page and select "manage topics."