Practical guide for running Claudex day-to-day — local models, cloud providers, diagnostics, and Telegram.
- CLI agent loop: reads/writes files, runs terminal commands, assists with coding
- Provider profile system (
profile:init,dev:profile,/providerin CLI) - Smart multi-provider router (
ROUTER_MODE=smart) - Runtime diagnostics (
doctor:runtime,doctor:report) - Telegram gateway (
claudex telegram) - Startup themes (
CLAUDEX_THEME) - Providers: NVIDIA AI, Gemini, OpenAI, Ollama, Atomic Chat, Codex, DeepSeek, Groq, and more
bun run dev:profileQuick preset switches:
bun run dev:fast # low latency (llama3.2:3b)
bun run dev:code # better coding quality (qwen2.5-coder:7b)
bun run dev:nvidia # NVIDIA Kimi K2# NVIDIA AI (free key)
bun run profile:init -- --provider nvidia --api-key nvapi-...
# Ollama (local)
bun run profile:init -- --provider ollama --model llama3.1:8b
# OpenAI
bun run profile:init -- --provider openai --api-key sk-... --model gpt-4o
# Gemini (free key)
bun run profile:init -- --provider gemini --api-key your-key
# Auto-select best local model for a goal
bun run profile:init -- --provider ollama --goal codingOr configure interactively from inside the CLI:
/provider
Get-Content .\.claudex-profile.jsonbun run doctor:runtimebun run profile:init -- --provider nvidia --api-key nvapi-...
bun run dev:nvidiaFree key at build.nvidia.com.
bun run profile:init -- --provider ollama --model llama3.1:8b
bun run dev:profileNo API key required. OPENAI_BASE_URL = http://localhost:11434/v1.
bun run profile:init -- --provider openai --api-key sk-... --model gpt-4o
bun run dev:openaibun run profile:init -- --provider gemini --api-key your-key
bun run dev:gemini$env:ROUTER_MODE="smart"
$env:ROUTER_STRATEGY="balanced"
bun run dev:profilebun run doctor:runtime # human-readable checks
bun run doctor:runtime:json # JSON output
bun run doctor:report # save to reports/doctor-runtime.json
bun run hardening:check # smoke + runtime doctor
bun run hardening:strict # typecheck + hardening| Goal | Model |
|---|---|
| Fast / general | llama3.1:8b |
| Better coding | qwen2.5-coder:14b |
| Low resource | llama3.2:3b |
| Best coding (if hardware allows) | qwen2.5-coder:32b |
Switch quickly:
bun run profile:init -- --provider ollama --model qwen2.5-coder:14b
bun run dev:profileGoal-based auto-selection:
bun run profile:init -- --provider ollama --goal latency
bun run profile:init -- --provider ollama --goal balanced
bun run profile:init -- --provider ollama --goal codingRun Claudex as a Telegram bot so you can use it from your phone.
# one-time setup
claudex telegram setup --token 123456:ABC --provider nvidia
# allow yourself (find your ID via @userinfobot on Telegram)
claudex telegram permit 987654321
# start the gateway
claudex telegram startOr from inside the CLI:
/telegram setup --token 123456:ABC --provider nvidia
/telegram permit 987654321
/telegram status
Full guide: telegram-gateway/README.md
$env:CLAUDEX_THEME="ocean" # sunset | ocean | aurora | neon | mono
claudexWrong folder. Navigate to the project root:
cd C:\path\to\claudex
bun run dev:profilewinget install Ollama.Ollama
# open a new terminal, then:
ollama --versionollama serve
# in another terminal:
bun run doctor:runtimeNo provider configured. Either:
bun run dev:profile # if you have a saved profileOr set env vars manually:
$env:CLAUDE_CODE_USE_NVIDIA="1"
$env:NVIDIA_API_KEY="nvapi-your-key"
claudex- "Map this repository architecture and explain the execution flow from entrypoint to tool invocation."
- "Find the top 5 risky modules and explain why."
- "Refactor this module for clarity without behavior change, then run checks and summarize diff impact."
- "Extract shared logic from duplicated functions and add minimal tests."
- "Reproduce the failure, identify root cause, implement fix, and validate with commands."
- "Trace this error path and list likely failure points with confidence levels."
- "Add runtime guardrails and fail-fast messages for invalid provider env vars."
- "Create a diagnostic command that outputs a JSON report for CI artifacts."
- "Do a code review of unstaged changes, prioritize bugs/regressions, and suggest concrete patches."
- Run
doctor:runtimebefore debugging provider issues - Prefer
dev:profileover manual env edits - Keep
.claudex-profile.jsonlocal (already gitignored) - Use
doctor:reportbefore asking for help — gives a reproducible snapshot
bun run doctor:runtime
bun run doctor:report
bun run smokeIf local model is slow:
ollama ps # PROCESSOR=CPU means valid but higher latency# profiles
bun run profile:init -- --provider nvidia --api-key nvapi-...
bun run profile:init -- --provider ollama --model llama3.1:8b
bun run profile:init -- --provider openai --api-key sk-... --model gpt-4o
bun run profile:init -- --provider gemini --api-key your-key
bun run profile:fast # preset: llama3.2:3b
bun run profile:code # preset: qwen2.5-coder:7b
bun run profile:nvidia # preset: kimi-k2-instruct
# launch
bun run dev:profile
bun run dev:nvidia
bun run dev:ollama
bun run dev:openai
bun run dev:gemini
bun run dev:codex
bun run dev:atomic-chat
bun run dev:fast
bun run dev:code
# diagnostics
bun run doctor:runtime
bun run doctor:runtime:json
bun run doctor:report
bun run smoke
bun run hardening:check
bun run hardening:strict
# telegram
claudex telegram setup --token <token> --provider <provider>
claudex telegram permit <user-id>
claudex telegram revoke <user-id>
claudex telegram status
claudex telegram startSetup is healthy when:
bun run doctor:runtimepasses all checksbun run dev:profileopens the CLI without errors- Model shown in the startup box matches your selected profile
- No Anthropic login screen appears (profile is saved)