The leading, most token-efficient MCP server for GitHub source code exploration via tree-sitter AST parsing
-
Updated
May 4, 2026 - Python
The leading, most token-efficient MCP server for GitHub source code exploration via tree-sitter AST parsing
Universal AI context generator. Saves thousands of tokens per conversation in Claude Code, Cursor, Copilot, Codex, and more.
SDL-MCP gives coding agents the right code context, not your entire repo. It turns sprawling codebases into compact, high-signal context that saves tokens, speeds up workflows, and improves agent output.
MCP server that saves Claude Code tokens by delegating bounded tasks to local or cloud LLMs. Works with LM Studio, Ollama, vLLM, DeepSeek, Groq, Cerebras.
MCP server for Claude Code and Codex. One tool call replaces ~42 minutes of agent exploration
Guardian Agent and Token Savings for Claude Code
A reversible code minifier for AI. Save tokens by stripping code format in your prompt, then perfectly restore it in the responces.
MCP server for Git with local Ollama — zero tokens for git operations
Turn any OpenAPI spec into a native CLI binary. No MCP, no bloat, no runtime dependencies, ONLY CLI.
TSCG — Deterministic tool-schema compiler for LLM agents. 50-72% token savings, 50 tools in 2.4ms. Phi-4 recovers from 0% to 90% accuracy. 459 tests, zero dependencies, MIT.
Caveman output style for Claude Code: 40% fewer output tokens, always-on formatting
Local-first Model Context Protocol (MCP) memory layer for Codex CLI/Desktop, Claude Code, Gemini CLI, Qwen/DeepSeek/Ollama and agent workflows. SQLite + FTS5 compact context packs, token savings, read-only mode, no external memory server.
Project-agnostic dual-memory MCP CLI for Claude Code, Cursor, and OpenCode (Qdrant tuned hybrid retrieval + structural memory hooks)
Make GitHub Copilot responses terse across VS Code Chat, Copilot CLI, cloud agent, and code review. One command. 40-75% fewer response tokens, no correctness hit.
Humanswith.ai MCP Stack: 21 local-first MCP servers for Claude Code, Codex, Cursor and Windsurf with repo retrieval, context prep, traces, screenshots and quality gates
Save Claude tokens by offloading tasks to free LLM APIs - 19x more workflows with the same token budget
Go single-binary persistent memory for AI coding agents. Saves 90-98% tokens vs re-reading files. CLI-first, full-text search, agent-agnostic.
Local Ollama-backed MCP server for Claude Code. Query large files without burning context (~30x token savings)
CodeCrew — Senior engineering team for your AI-generated code. One command. Saves $20-40/day in tokens.
lowfat - slim your command output. strips noise, saves tokens.
Add a description, image, and links to the token-savings topic page so that developers can more easily learn about it.
To associate your repository with the token-savings topic, visit your repo's landing page and select "manage topics."