Documentation Index
Fetch the complete documentation index at: https://mintlify.com/helicone/helicone/llms.txt
Use this file to discover all available pages before exploring further.
Overview
The Models endpoint returns a comprehensive list of all AI models supported by Helicone AI Gateway. This includes models from providers like OpenAI, Anthropic, Google, Meta, Mistral, and many others. The response follows the OpenAI API format for compatibility with existing tools and libraries.Endpoint
Authentication
No authentication is required for this endpoint. It’s publicly accessible to help you discover available models.Request
This endpoint accepts no parameters. Simply make a GET request:Response Format
Object type, always
"list"Array of model objects. Each model contains:
Model Object
The model identifier that can be used in API requests.Examples:
"gpt-4""claude-3-5-sonnet-20241022""gemini-2.0-flash-exp"
Object type, always
"model"Unix timestamp of when the model was released
The organization that owns/created the model.Examples:
"openai""anthropic""google""meta"
Example Request
Example Response
Model Categories
The endpoint returns models across multiple categories:Chat Models
- OpenAI GPT models (GPT-4, GPT-3.5, etc.)
- Anthropic Claude models
- Google Gemini models
- Meta Llama models
- Mistral models
- DeepSeek models
- And many more
Multimodal Models
Models that support multiple input/output types (text, images, audio):- GPT-4 Vision
- Claude 3 Opus/Sonnet
- Gemini Pro Vision
Specialized Models
- Reasoning models (o1, o3-mini)
- Fast inference models (GPT-4 Turbo, Gemini Flash)
- Cost-effective models (GPT-3.5, Claude Haiku)
Filtering Models
The API returns all publicly available models that don’t require explicit routing. Some enterprise or specialized models may require additional configuration and won’t appear in this list. Models requiring explicit routing (viarequireExplicitRouting flag) are excluded from the response to avoid confusion.
Usage with Chat Completions
Once you have the model ID from this endpoint, use it directly in the Chat Completions endpoint:Model Information
For detailed information about each model including:- Pricing
- Context length
- Capabilities
- Provider documentation
Programmatic Access
You can programmatically fetch and filter models in your application:Notes
- This endpoint is unauthenticated and rate-limited per IP address
- The list of models is updated automatically as new providers and models are added
- Some models may require specific authentication or configuration to use
- Model availability may vary by region or account type
- The
createdtimestamp uses Unix epoch time (seconds since January 1, 1970)
Related Endpoints
- Chat Completions - Use models to generate completions
- Supported Models - Browse models with detailed information
