AI Models

Access multiple AI models through unified endpoints. SolCognia aggregates leading AI models including GPT, Claude, and Gemini.

Available Models

List Models

Get information about all available AI models:

GET /api/v1/models

Response:

{
  "models": [
    {
      "id": "gpt-4",
      "name": "GPT-4",
      "provider": "OpenAI",
      "capabilities": ["chat", "completion", "code"],
      "max_tokens": 8192,
      "cost_per_token": 0.03,
      "available": true
    },
    {
      "id": "claude-3",
      "name": "Claude 3",
      "provider": "Anthropic",
      "capabilities": ["chat", "analysis", "writing"],
      "max_tokens": 100000,
      "cost_per_token": 0.025,
      "available": true
    }
  ]
}

Chat Completions

Create Chat Completion

Send messages to an AI model and receive responses:

Request Body:

Response:

Streaming Responses

For real-time responses, use streaming:

Streaming Response:

Text Completions

Create Completion

Generate text completions for prompts:

Request Body:

Response:

Model-Specific Features

GPT Models

Claude Models

Gemini Models

Model Comparison

Compare Responses

Get responses from multiple models for the same prompt:

Request Body:

Response:

Usage Analytics

Get Model Usage

Track your usage across different models:

Response:

Error Handling

Common Errors

Model Not Available (503)

Token Limit Exceeded (400)

Rate Limit Exceeded (429)

Best Practices

Optimize Token Usage

  • Use appropriate max_tokens limits

  • Implement response caching for repeated queries

  • Choose the right model for your use case

Handle Streaming

Monitor Performance

  • Track response times and success rates

  • Monitor token usage and costs

  • Set up alerts for unusual patterns

The Models API will be available with our Q4 2025 developer ecosystem launch.

Last updated