AI Models
Access multiple AI models through unified endpoints. SolCognia aggregates leading AI models including GPT, Claude, and Gemini.
Available Models
List Models
Get information about all available AI models:
GET /api/v1/modelsResponse:
{
"models": [
{
"id": "gpt-4",
"name": "GPT-4",
"provider": "OpenAI",
"capabilities": ["chat", "completion", "code"],
"max_tokens": 8192,
"cost_per_token": 0.03,
"available": true
},
{
"id": "claude-3",
"name": "Claude 3",
"provider": "Anthropic",
"capabilities": ["chat", "analysis", "writing"],
"max_tokens": 100000,
"cost_per_token": 0.025,
"available": true
}
]
}Chat Completions
Create Chat Completion
Send messages to an AI model and receive responses:
Request Body:
Response:
Streaming Responses
For real-time responses, use streaming:
Streaming Response:
Text Completions
Create Completion
Generate text completions for prompts:
Request Body:
Response:
Model-Specific Features
GPT Models
Claude Models
Gemini Models
Model Comparison
Compare Responses
Get responses from multiple models for the same prompt:
Request Body:
Response:
Usage Analytics
Get Model Usage
Track your usage across different models:
Response:
Error Handling
Common Errors
Model Not Available (503)
Token Limit Exceeded (400)
Rate Limit Exceeded (429)
Best Practices
Optimize Token Usage
Use appropriate
max_tokenslimitsImplement response caching for repeated queries
Choose the right model for your use case
Handle Streaming
Monitor Performance
Track response times and success rates
Monitor token usage and costs
Set up alerts for unusual patterns
The Models API will be available with our Q4 2025 developer ecosystem launch.
Last updated