LLM Proxy
LLM Proxy API docs
Test Authentication
Paste a valid LLM Proxy API key to run endpoint tests.
Chat Completions
Primary text generation endpoint compatible with OpenAI chat clients.
POST/v1/chat/completions
cURL example
bash
curl -s http://localhost:3000/v1/chat/completions \
-H "Authorization: Bearer $LLM_PROXY_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-4.1-mini",
"messages": [{"role":"user","content":"Say hello from LLM Proxy"}]
}'Request body example
json
{
"model": "gpt-4.1-mini",
"messages": [
{ "role": "system", "content": "You are a helpful assistant." },
{ "role": "user", "content": "Write one short product tagline for an AI proxy." }
],
"temperature": 0.7,
"max_tokens": 120
}Response example
json
{
"id": "chatcmpl_123",
"object": "chat.completion",
"created": 1735678910,
"model": "gpt-4.1-mini",
"choices": [
{
"index": 0,
"message": {
"role": "assistant",
"content": "One API key, every model, zero integration chaos."
},
"finish_reason": "stop"
}
],
"usage": {
"prompt_tokens": 22,
"completion_tokens": 14,
"total_tokens": 36
}
}Try It
POST
/v1/chat/completionsRun a live request using a pasted API key.
Select an account key or paste a manual key first to test endpoints.