LLM Proxy
LLM Proxy API docs
Test Authentication
Paste a valid LLM Proxy API key to run endpoint tests.

Chat Completions

Primary text generation endpoint compatible with OpenAI chat clients.

POST/v1/chat/completions

cURL example

bash
curl -s http://localhost:3000/v1/chat/completions \
  -H "Authorization: Bearer $LLM_PROXY_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "gpt-4.1-mini",
    "messages": [{"role":"user","content":"Say hello from LLM Proxy"}]
  }'

Request body example

json
{
  "model": "gpt-4.1-mini",
  "messages": [
    { "role": "system", "content": "You are a helpful assistant." },
    { "role": "user", "content": "Write one short product tagline for an AI proxy." }
  ],
  "temperature": 0.7,
  "max_tokens": 120
}

Response example

json
{
  "id": "chatcmpl_123",
  "object": "chat.completion",
  "created": 1735678910,
  "model": "gpt-4.1-mini",
  "choices": [
    {
      "index": 0,
      "message": {
        "role": "assistant",
        "content": "One API key, every model, zero integration chaos."
      },
      "finish_reason": "stop"
    }
  ],
  "usage": {
    "prompt_tokens": 22,
    "completion_tokens": 14,
    "total_tokens": 36
  }
}
Try It
POST
/v1/chat/completions
Run a live request using a pasted API key.

Select an account key or paste a manual key first to test endpoints.