Models
One API for hundreds of models - explore 340+ models and providers
Models
One API for hundreds of models
Explore and browse 340+ models and providers through our unified API. Access models from OpenAI, Anthropic, Google, Meta, Mistral, DeepSeek, and dozens more providers—all with a single API key.
View the complete list of all models with live pricing at AI Gateway Pricing.
Models API
Our Models API makes the most important information about all LLMs freely available. Query model metadata, pricing, capabilities, and supported parameters programmatically.
Endpoint
Example Request
API Response Schema
The Models API returns a standardized JSON response format that provides comprehensive metadata for each available model. This schema is designed for reliable integration with production applications.
Root Response Object
Model Object Schema
Each model in the data array contains the following standardized fields:
| Field | Type | Description |
|---|---|---|
id | string | Unique model identifier used in API requests (e.g., "openai/gpt-4o") |
name | string | Human-readable display name for the model |
created | number | Unix timestamp of when the model was added |
description | string | Detailed description of the model's capabilities |
context_length | number | Maximum context window size in tokens |
architecture | Architecture | Object describing the model's technical capabilities |
pricing | Pricing | Price structure for using this model |
top_provider | TopProvider | Configuration details for the primary provider |
per_request_limits | object | null | Rate limiting information (null if no limits) |
supported_parameters | string[] | Array of supported API parameters for this model |
Example Model Object
Architecture Object
Describes the model's technical capabilities and supported modalities.
| Field | Type | Description |
|---|---|---|
input_modalities | string[] | Supported input types: "text", "image", "file", "audio" |
output_modalities | string[] | Supported output types: "text", "image", "audio" |
tokenizer | string | Tokenization method used by the model |
instruct_type | string | null | Instruction format type ("chat", "completion", or null) |
Input Modalities
| Modality | Description |
|---|---|
text | Standard text input (all models) |
image | Image/vision input (GPT-4o, Claude 3, Gemini) |
file | Document/file input (some models) |
audio | Audio input (GPT-4o Audio, Gemini) |
Pricing Object
All pricing values are in USD per token. A value of "0" indicates the feature is free.
For current pricing on all models, visit AI Gateway Pricing.
| Field | Type | Description |
|---|---|---|
prompt | string | Cost per input token |
completion | string | Cost per output token |
request | string | Fixed cost per API request |
image | string | Cost per image input |
input_cache_read | string | Cost per cached input token read |
input_cache_write | string | Cost per cached input token write |
Calculating Costs
Top Provider Object
Configuration details for the primary provider serving this model.
| Field | Type | Description |
|---|---|---|
context_length | number | Provider-specific context limit |
max_completion_tokens | number | Maximum tokens in response |
is_moderated | boolean | Whether content moderation is applied |
Supported Parameters
The supported_parameters array indicates which OpenAI-compatible parameters work with each model:
| Parameter | Description |
|---|---|
tools | Function calling capabilities |
tool_choice | Tool selection control |
max_tokens | Response length limiting |
temperature | Randomness control (0-2) |
top_p | Nucleus sampling threshold |
response_format | Output format specification (e.g., JSON mode) |
stop | Custom stop sequences |
frequency_penalty | Repetition reduction (-2 to 2) |
presence_penalty | Topic diversity (-2 to 2) |
seed | Deterministic outputs |
structured_outputs | JSON schema enforcement |
Checking Parameter Support
Token Counts and Pricing
Different models tokenize text in different ways. Some models break text into chunks of multiple characters (GPT, Claude, Llama), while others tokenize differently. Token counts and costs vary between models even with identical inputs.
Costs are calculated according to each model's tokenizer. Use the usage field in API responses to get accurate token counts:
Filtering Models
Query specific models or filter by capabilities:
Model Updates
Models and pricing are updated regularly. For the latest information:
- Browse models visually: AI Gateway Pricing
- Query programmatically: Use the
/v1/modelsendpoint