Intelligent LLM model router driven by real code metrics.
README.md— project overview, install, and quickstartdocs/README.md— generated API inventory from source analysisdocs/llx-tools.md— ecosystem CLI referencedocs/PRIVACY.md— anonymization and sensitive-data handling
Successor to preLLM — rebuilt with modular architecture, no god modules, and metric-driven routing.
llx analyzes your codebase with code2llm, redup, and vallm, then selects the optimal LLM model based on actual project metrics — file count, complexity, coupling, duplication — not abstract scores.
Principle: larger + more coupled + more complex → stronger (and more expensive) model.
llx is organized around a small set of command groups:
llx analyze,llx select,llx chat— metric-driven analysis and model routingllx proxy— LiteLLM proxy config, start, and statusllx mcp— MCP server start, config, and tool listingllx plan— planfile generation, review, code generation, and executionllx strategy— interactive strategy creation, validation, run, and verificationllx info,llx models,llx init,llx fix— inspection and utility commands
preLLM proved the concept but had architectural issues that llx resolves:
| Problem in preLLM | llx Solution |
|---|---|
cli.py: 999 lines, CC=30 (main), CC=27 (query) |
CLI split into app.py + formatters.py, max CC ≤ 8 |
core.py: 893 lines god module |
Config, analysis, routing in separate modules (≤250L each) |
trace.py: 509 lines, CC=28 (to_stdout) |
Output formatting as dedicated functions |
| Hardcoded model selection | Metric-driven thresholds from code2llm .toon data |
| No duplication/validation awareness | Integrates redup + vallm for richer metrics |
┌─────────────────────────────────────────────────────────────┐
│ IDE / Agent Layer │
│ Roo Code │ Cline │ Continue.dev │ Aider │ Claude Code │
│ (point at localhost:4000 as OpenAI-compatible API) │
└─────────────────┬───────────────────────────────────────────┘
│
┌─────────────────▼───────────────────────────────────────────┐
│ LiteLLM Proxy (localhost:4000) │
│ ┌──────────┐ ┌──────────────┐ ┌────────────────────┐ │
│ │ Router │ │ Semantic │ │ Cost Tracking │ │
│ │ (metrics)│ │ Cache (Redis)│ │ + Budget Limits │ │
│ └────┬─────┘ └──────────────┘ └────────────────────┘ │
└───────┼─────────────────────────────────────────────────────┘
│
┌────┼────────────────────────────────────────┐
│ │ Model Tiers │
│ ├── premium: Claude Opus 4 │
│ ├── balanced: Claude Sonnet 4 / GPT-5 │
│ ├── cheap: Claude Haiku 4.5 │
│ ├── free: Gemini 2.5 Pro │
│ ├── openrouter: 300+ models (fallback) │
│ └── local: Ollama (Qwen2.5-Coder) │
└──────────────────────────────────────────────┘
│
┌───────▼─────────────────────────────────────────────────────┐
│ Code Analysis Pipeline │
│ code2llm → redup → vallm → llx │
│ (metrics → duplication → validation → model selection) │
└─────────────────────────────────────────────────────────────┘
llx exposes its MCP tools through a shared registry in llx.mcp.tools.MCP_TOOLS.
By default, the MCP server runs over stdio for Claude Desktop. Use SSE only when you need a remote or web client.
# Start MCP server for Claude Desktop (stdio)
llx mcp start
# Start MCP server over SSE for web/remote clients
llx mcp start --mode sse --port 8000
# Generate Claude Desktop config
llx mcp config
# List the live MCP registry
llx mcp tools
# Direct module entrypoint
python -m llx.mcp --sse --port 8000llx_analyze,llx_select,llx_chat— project metrics and model routingllx_preprocess,llx_context— query preprocessing and environment contextcode2llm_analyze,redup_scan,vallm_validate— code-quality analysis helpersllx_proxy_status,llx_proxym_status,llx_proxym_chat— proxy and proxym integrationaider,planfile_generate,planfile_apply— workflow and refactoring helpersllx_privacy_scan,llx_project_anonymize,llx_project_deanonymize— privacy tooling
{
"mcpServers": {
"llx": {
"command": "python3",
"args": ["-m", "llx.mcp"]
}
}
}pip install llx
# With integrations
pip install llx[all] # Everything + MCP
pip install llx[mcp] # MCP server only
pip install llx[litellm] # LiteLLM proxy
pip install llx[code2llm] # Code analysis
pip install llx[redup] # Duplication detection
pip install llx[vallm] # Code validation# Analyze project and get model recommendation
llx analyze ./my-project
# Quick model selection
llx select .
# With task hint
llx select . --task refactor
# Point to pre-existing .toon files
llx analyze . --toon-dir ./analysis/
# JSON output for CI/CD
llx analyze . --json
# Chat with auto-selected model
llx chat . --prompt "Refactor the god modules"
# Force local model
llx select . --local| Metric | Premium (≥) | Balanced (≥) | Cheap (≥) | Free |
|---|---|---|---|---|
| Files | 50 | 10 | 3 | <3 |
| Lines | 20,000 | 5,000 | 500 | <500 |
| Avg CC | 6.0 | 4.0 | 2.0 | <2.0 |
| Max fan-out | 30 | 10 | — | — |
| Max CC | 25 | 15 | — | — |
| Dup groups | 15 | 5 | — | — |
| Dep cycles | any | — | — | — |
LLX provides reversible anonymization to protect sensitive data when sending to LLMs:
- Text anonymization: Emails, API keys, passwords, PESEL, credit cards
- Project-level: AST-based code anonymization (variables, functions, classes)
- Round-trip: Anonymize → Send to LLM → Deanonymize response
- Persistent mapping: Save/restore context for later deanonymization
from llx.privacy import quick_anonymize, quick_deanonymize
# Simple text anonymization
result = quick_anonymize("Email: user@example.com, API: sk-abc123")
print(result.text) # "Email: [EMAIL_A1B2], API: [APIKEY_C3D4]"
# Later: restore original values
restored = quick_deanonymize(llm_response, result.mapping)from llx.privacy.project import AnonymizationContext, ProjectAnonymizer
from llx.privacy.deanonymize import ProjectDeanonymizer
# Anonymize entire project
ctx = AnonymizationContext(project_path="./my-project")
anonymizer = ProjectAnonymizer(ctx)
result = anonymizer.anonymize_project()
# Save context for later
ctx.save("./my-project.anon.json")
# Deanonymize LLM response
deanonymizer = ProjectDeanonymizer(ctx)
restored = deanonymizer.deanonymize_chat_response(llm_response)// Scan for sensitive data
{"tool": "llx_privacy_scan", "text": "Email: user@example.com"}
// Anonymize project
{"tool": "llx_project_anonymize", "path": "./my-project", "output_dir": "./anon"}
// Deanonymize response
{"tool": "llx_project_deanonymize", "context_path": "./anon/.anonymization_context.json", "text": "Fix fn_ABC123"}See docs/PRIVACY.md and examples/privacy/ for complete documentation.
| Project | Files | Lines | CC̄ | Max CC | Fan-out | Tier |
|---|---|---|---|---|---|---|
| Single script | 1 | 80 | 2.0 | 4 | 0 | free |
| Small CLI | 5 | 600 | 3.0 | 8 | 3 | cheap |
| preLLM | 31 | 8,900 | 5.0 | 28 | 30 | premium |
| vallm | 56 | 8,604 | 3.5 | 42 | — | balanced |
| code2llm | 113 | 21,128 | 4.6 | 65 | 45 | premium |
| Monorepo | 500+ | 100K+ | 5.0+ | 30+ | 50+ | premium |
llx proxy config # Generate litellm_config.yaml
llx proxy start # Start proxy on :4000
llx proxy status # Check if runningConfigure IDE tools to point at http://localhost:4000:
| Tool | Config |
|---|---|
| Roo Code / Cline | "apiBase": "http://localhost:4000/v1" |
| Continue.dev | "apiBase": "http://localhost:4000/v1" |
| Aider | OPENAI_API_BASE=http://localhost:4000 |
| Claude Code | ANTHROPIC_BASE_URL=http://localhost:4000 |
| Cursor / Windsurf | OpenAI-compatible endpoint |
llx init # Creates llx.toml with defaultsEnvironment variables: LLX_LITELLM_URL, LLX_DEFAULT_TIER, LLX_PROXY_PORT, LLX_VERBOSE.
from llx import analyze_project, select_model, LlxConfig
metrics = analyze_project("./my-project")
result = select_model(metrics)
print(result.model_id) # "claude-opus-4-20250514"
print(result.explain()) # Human-readable reasoning| Tool | Role | llx Uses |
|---|---|---|
| code2llm | Static analysis | CC, fan-out, cycles, hotspots |
| redup | Duplication detection | Groups, recoverable lines |
| vallm | Code validation | Pass rate, issue count |
| llx | Model routing + MCP server | Consumes all above |
llx/
├── __init__.py
├── config.py
├── analysis/ # Project metrics and external tool runners
├── cli/ # Typer commands and terminal formatters
├── commands/ # High-level command helpers
├── detection/ # Project type detection
├── integrations/ # Proxy, proxym, and context helpers
├── mcp/ # MCP server, client, service, and tool registry
├── orchestration/ # Multi-instance coordination utilities
├── planfile/ # Strategy generation and execution helpers
├── prellm/ # Small→large LLM preprocessing pipeline
├── privacy/ # Anonymization and deanonymization helpers
├── routing/ # Model selection and LiteLLM client
└── tools/ # Docker, VS Code, models, config, health utilities
Full generated API inventory: docs/README.md.
- Shared MCP registry:
llx.mcp.tools.MCP_TOOLSpowers bothllx mcp toolsand the server dispatcher. - Single tier order:
routing/selector.pyuses oneTIER_ORDERconstant for selection and context-window upgrades. - Version alignment: the package exports now match
pyproject.tomlandVERSION. - Focused modules: CLI, routing, analysis, integrations, and planfile code are split by responsibility.
Licensed under Apache-2.0.
Tom Sapletta
