AI Provider Connectors
DgiDgi supports 25+ AI providers for LLM inference, giving you flexibility to choose the best model for your use case, budget, and compliance requirements.
Supported Providers
Tier 1: Major Cloud Providers
| Provider | Models | Key Features | Pricing |
|---|---|---|---|
| OpenAI | GPT-4o, GPT-4o-mini, o1-preview | Industry standard, vision, tools | Pay-per-token |
| Anthropic | Claude 3.5 Sonnet, Claude 3 Haiku | 200K context, safety-focused | Pay-per-token |
| Google Gemini | Gemini 1.5 Pro, Gemini 1.5 Flash | 2M context, multimodal | Pay-per-token |
| Azure OpenAI | GPT-4o (Azure), GPT-4o-mini (Azure) | Enterprise compliance, SLA | Pay-per-token |
| AWS Bedrock | Claude, Llama, Titan, Cohere | AWS integration, regulated | Pay-per-token |
Tier 2: High-Performance Inference
| Provider | Models | Key Features | Pricing |
|---|---|---|---|
| Groq | Llama 3.3 70B, Mixtral 8x7B | Ultra-fast LPU inference | Pay-per-token |
| Cerebras | Llama 3.3 70B, Llama 3.1 8B | 2000+ tokens/sec | Pay-per-token |
| SambaNova | Llama 3.3 70B, DeepSeek R1, Qwen | Enterprise-grade | Pay-per-token |
| Fireworks | Llama 3.3 70B, DeepSeek V3 | Fast, cost-effective | Pay-per-token |
| Lepton | Llama 3.1 405B, Llama 3.3 70B | Serverless GPU | Pay-per-token |
Tier 3: Specialized Providers
| Provider | Models | Key Features | Pricing |
|---|---|---|---|
| Mistral | Mistral Large, Codestral | European, coding-focused | Pay-per-token |
| Cohere | Command R+ | Enterprise embeddings, RAG | Pay-per-token |
| Perplexity | Sonar Large | Search-augmented, real-time | Pay-per-token |
| xAI | Grok 2, Grok Vision | Multimodal, reasoning | Pay-per-token |
| DeepSeek | DeepSeek V3, DeepSeek Coder | Cost-effective, coding | Pay-per-token |
Tier 4: Aggregators & Open Source
| Provider | Models | Key Features | Pricing |
|---|---|---|---|
| OpenRouter | 100+ models | Unified API, fallback | Pay-per-token |
| Together AI | Llama, Qwen, Mixtral | Training + inference | Pay-per-token |
| Replicate | Llama, Mistral, custom | Model marketplace | Pay-per-use |
| Hugging Face | Open-source models | Community models | Free tier |
| GitHub Models | Copilot Enterprise/Pro | GitHub integration | Subscription |
Tier 5: Self-Hosted
| Provider | Models | Key Features | Pricing |
|---|---|---|---|
| Ollama | Llama, Qwen, DeepSeek | Local, private, FREE | Free |
| DgiDgi Self-Hosted | Any OpenAI-compatible | Your infrastructure | Your cost |
Model Catalog
Fast Tier (Low Latency)
Best for interactive chat, quick responses:
- GPT-4o-mini (OpenAI)
- Claude 3 Haiku (Anthropic)
- Gemini 1.5 Flash (Google)
- Llama 3.3 70B (Groq, Cerebras)
- Mixtral 8x7B (Groq)
Balanced Tier (Cost/Performance)
Best for general tasks:
- GPT-4o (OpenAI)
- Claude 3.5 Sonnet (Anthropic)
- Gemini 1.5 Pro (Google)
- Mistral Large (Mistral)
- DeepSeek V3 (DeepSeek)
Powerful Tier (Maximum Capability)
Best for complex reasoning, code generation:
- o1-preview (OpenAI)
- Claude 3.5 Sonnet (Anthropic)
- Llama 3.1 405B (Together, Replicate, Lepton)
- Grok 2 (xAI)
Reasoning Tier (Extended Thinking)
Best for multi-step problems:
- o1-preview (OpenAI)
- DeepSeek R1 Distill (SambaNova)
Provider Configuration
Platform-Level (Admin)
Configure default providers for all tenants:
# Environment variables
OPENAI_API_KEY=sk-...
ANTHROPIC_API_KEY=sk-ant-...
GROQ_API_KEY=gsk_...
GEMINI_API_KEY=...
Tenant-Level (User)
Users can add their own API keys via Settings > Connectors > AI.
Routing & Fallback
DgiDgi's LLM Gateway automatically routes requests based on:
- Task Type: Code generation, chat, analysis, etc.
- Required Capabilities: Vision, tools, reasoning
- Cost Optimization: Tenant budgets, token limits
- Availability: Provider health, rate limits
Fallback chain example:
Primary: OpenAI GPT-4o
-> Fallback 1: Anthropic Claude 3.5 Sonnet
-> Fallback 2: Groq Llama 3.3 70B
-> Fallback 3: Gemini 1.5 Pro
Cost Management
Per-Tenant Budgets
Set daily/monthly limits per tenant to control costs.
Model Tiers
- Fast: ~$0.0001-0.001/1K tokens
- Balanced: ~$0.001-0.005/1K tokens
- Powerful: ~$0.005-0.015/1K tokens
- Reasoning: ~$0.015-0.06/1K tokens
Free Options
- Ollama (local, requires hardware)
- Hugging Face free tier
- DgiDgi self-hosted (your infrastructure)
Self-Hosted LLMs (DgiDgi Provider)
Deploy your own LLMs and connect them to DgiDgi:
- Deploy an OpenAI-compatible API server (vLLM, text-generation-inference, etc.)
- Configure the DgiDgi provider:
- Base URL: Your LLM server endpoint (e.g.,
https://your-llm.example.com/v1) - API Key: Optional authentication key
- Base URL: Your LLM server endpoint (e.g.,
- Models will be available to all users in your tenant
Supported frameworks:
- vLLM
- text-generation-inference (TGI)
- llama.cpp server
- Ollama (remote)
- Any OpenAI-compatible API