Supported LLM Platforms
PipeLLM Gateway supports mainstream LLM service providers. Biggest advantage: Use official SDKs (Anthropic, OpenAI, Google, etc.), just change baseURL to access any platform transparently. No need to learn new APIs, zero code changes! Complete model list: https://www.pipellm.com/models🤖 OpenAI Ecosystem
OpenAI Official
Status: ✅ Fully supported Usage: Use OpenAI official SDK- GPT-4 Series:
gpt-4,gpt-4-turbo,gpt-4o,gpt-4o-mini - GPT-3.5 Series:
gpt-3.5-turbo,gpt-3.5-turbo-16k - Embeddings:
text-embedding-ada-002,text-embedding-3-small,text-embedding-3-large - Speech:
whisper-1 - Images:
dall-e-3,dall-e-2
Azure OpenAI
Status: ✅ Fully supported Usage: Use OpenAI SDK, transparently call Azure services- GPT-4 Series:
gpt-4,gpt-4-32k,gpt-4-turbo,gpt-4o - GPT-3.5 Series:
gpt-35-turbo,gpt-35-turbo-16k - Embeddings:
text-embedding-ada-002
🦙 Anthropic Claude
Status: ✅ Fully supported Usage: Use Anthropic official SDK- Claude 3 Series:
claude-3-haiku- Fast, economicalclaude-3-sonnet- Balanced performanceclaude-3-opus- Highest qualityclaude-3-5-sonnet- Latest version
- Long context processing (supports 200K tokens)
- Excellent reasoning and analysis
- Strong instruction following
- Tool usage support
| Your SDK | Actual Platform | Notes |
|---|---|---|
| Anthropic SDK | AWS Bedrock | Auto converts to Bedrock protocol |
| Anthropic SDK | Google Vertex | Auto converts to Vertex protocol |
| Anthropic SDK | Azure | Auto converts to Azure protocol |
| Anthropic SDK | Anthropic Official | Direct call |
🤖 Google Gemini
Status: ✅ Fully supported Usage: Use Google’s native libraries or standard APIgemini-pro- General-purpose modelgemini-pro-vision- Multimodal modelgemini-ultra- Advanced model (if available)
- Multimodal capabilities (text + images)
- Code generation optimization
- Fast response
- Google ecosystem integration
☁️ AWS Bedrock
Status: ✅ Fully supported Usage: Use native SDKs to call Bedrock services- Anthropic:
claude-3-haiku,claude-3-sonnet,claude-3-opus - Amazon Titan:
amazon.titan-text-express-v1,amazon.titan-text-lite-v1 - AI21 Labs:
ai21.j2-mid,ai21.j2-ultra - Cohere:
cohere.command-text-v14,cohere.command-light-text-v14 - Meta:
meta.llama3-8b-instruct,meta.llama3-70b-instruct,meta.llama3-1-8b-instruct,meta.llama3-1-70b-instruct,meta.llama3-2-11b-vision,meta.llama3-2-90b-vision - Mistral:
mistral.mistral-7b-instruct,mistral.mixtral-8x7b-instruct,mistral.mistral-large-latest,mistral.mistral-small-latest - Stability AI:
stability.stable-diffusion-xl-v1
- AWS native integration
- Enterprise-level security
- Scalability
- Pay-per-use
🌐 Other Cloud Platforms
Google Vertex AI
Status: ✅ Fully supportedFireworks AI
Status: ✅ Fully supported Supported models:accounts/fireworks/models/firefunction-v2accounts/fireworks/models/llama-v3p1-405baccounts/fireworks/models/llama-v3p1-70baccounts/fireworks/models/llama-v3p1-8baccounts/stabilityai/models/stable-diffusion-xl-1024-v1-0- Other Fireworks models
Together AI
Status: ✅ Fully supported Supported models:Nous-Hermes-2-Mixtral-8x7B-DPOLlama-3-8b-SFTCodeLlama-34bWizardLM-2-8x22b- Other Together AI models
Groq
Status: ✅ Fully supported Supported models:llama3-8b-8192llama3-70b-8192mixtral-8x7b-32768gemma-7b-itgemma2-9b-itllama-3-3-70b-versatilellama-3-3-8b-instant
- Ultra-fast inference
- Low latency
- Real-time application optimization
Replicate
Status: ✅ Fully supported Supported features:- Image generation
- Video generation
- Audio processing
- Custom model deployment
OpenRouter
Status: ✅ Fully supported Features:- Aggregate multiple providers
- Unified billing
- Simplified access
🎨 Media Processing Platforms
Stability AI
Status: ✅ Fully supported Supported models:stable-diffusion-xl-1024-v1-0stable-diffusion-3stable-cascadestable-video-diffusion
Ideogram
Status: ✅ Fully supported Features:- Creative image generation
- Text rendering optimization
- Artistic styles
Luma Labs
Status: ✅ Fully supported Supported features:- 3D model generation
- Image-to-3D conversion
- Video processing
📊 Platform Comparison
| Feature | OpenAI | Anthropic | Gemini | AWS Bedrock | Azure |
|---|---|---|---|---|---|
| Max Context | 128K | 200K | 32K | 200K | 128K |
| Multimodal | ✅ | ❌ | ✅ | Partial | Partial |
| Code Ability | ⭐⭐⭐⭐⭐ | ⭐⭐⭐⭐ | ⭐⭐⭐⭐ | ⭐⭐⭐ | ⭐⭐⭐⭐ |
| Reasoning | ⭐⭐⭐⭐ | ⭐⭐⭐⭐⭐ | ⭐⭐⭐⭐ | ⭐⭐⭐⭐ | ⭐⭐⭐⭐ |
| Speed | Fast | Very Fast | Very Fast | Fast | Fast |
| Price | High | Medium | Low | Medium | Medium |
| Enterprise | ✅ | ✅ | ✅ | ✅ | ✅ |
🚀 How to Choose the Right Platform
1. By Use Case
Code Generation:- Best: OpenAI GPT-4o, Claude 3
- Features: High accuracy, multi-language support
- Best: Claude 3 (200K context)
- Features: Can process entire books or codebases
- Best: OpenAI GPT-4o, Gemini Pro
- Features: High creativity, diverse styles
- Best: Azure OpenAI, AWS Bedrock
- Features: Enterprise security, data guarantees
- Best: Gemini Pro, Llama 3
- Features: High cost-effectiveness
2. By Technical Requirements
Multimodal Needs:- OpenAI GPT-4o
- Google Gemini Pro Vision
- AWS Titan Multimodal
- Anthropic Claude 3 (200K)
- OpenAI GPT-4o (128K)
- AWS Claude 3 (200K)
- Groq (hardware acceleration)
- OpenAI (optimized network)
- Anthropic (fast models)
3. Using PipeLLM Gateway Advantages
No Manual Selection:- Auto-select based on load
- Route based on model availability
- Optimize based on cost
- Auto-switch if provider unavailable
- Ensure service continuity
- Reduce downtime risk
🛠️ Advanced Configuration
1. Specify Provider Preference
2. Force Specific Provider
3. Model Aliases
4. Region Selection
📈 Performance Monitoring
Monitor via management dashboard:- Success rate comparison across providers
- Average response time
- Cost analysis
- Model usage statistics
🤝 Support
If you need platform support or encounter issues:- Check documentation: Visit provider’s official docs
- Enable debug mode: Use
X-Debug: trueto view details - Contact support: Email [email protected]
Tip: PipeLLM Gateway continuously adds support for new platforms and models. Stay updated!