Quickstart
1) Build CLI
go build -o llm-proxy ./cmd/proxy
2) Setup
./llm-proxy setup --interactive
3) Start services
./llm-proxy server
./llm-proxy admin --management-token $MANAGEMENT_TOKEN
3b) Start with Redis caching (optional)
# Start Redis
docker run -d --name redis -p 6379:6379 redis:alpine
# Start proxy with caching
HTTP_CACHE_ENABLED=true HTTP_CACHE_BACKEND=redis REDIS_ADDR=localhost:6379 ./llm-proxy server
4) Generate a token
- In the Admin UI: Tokens → Generate Token (recommended)
- Or via CLI (replace IDs):
llm-proxy manage token generate --project-id <project-id> --management-token $MANAGEMENT_TOKEN --duration 1440
5) Test chat via proxy (interactive)
./llm-proxy openai chat --token $PROXY_TOKEN --model gpt-4.1-mini
6) Test API via curl
curl -sS -X POST http://localhost:8080/v1/chat/completions \
-H "Authorization: Bearer $PROXY_TOKEN" \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-4.1-mini",
"messages": [
{"role":"user","content":"Hello!"}
]
}'
See CLI reference and Architecture.