Base URL
All API requests are made to:OpenAI-Compatible Format
ModelStack follows the OpenAI API format, so you can use any existing OpenAI SDK or HTTP client. Just change the base URL and API key.Available Endpoints
| Method | Endpoint | Description |
|---|---|---|
POST | /v1/chat/completions | Create a chat completion |
GET | /v1/models | List available models |
Authentication
All requests require an API key passed via theAuthorization header:
Request Format
All requests use JSON. Set theContent-Type header:
Response Format
Responses follow the OpenAI response format:Streaming
ModelStack supports Server-Sent Events (SSE) streaming. Set"stream": true in your request body to receive incremental responses.
Rate Limits
Rate limits depend on your subscription plan:| Plan | Rate Limit |
|---|---|
| Starter | 15 RPM |
| Pro | 40 RPM |
| Max | 60 RPM |
429 Too Many Requests error.