# Model Providers
URL: /models
import { Card, Cards } from "fumadocs-ui/components/card";
Tambo supports multiple AI model providers, allowing you to choose the best model for your application's needs. Each provider offers different models with varying capabilities, pricing, and performance characteristics.
## Available Providers
Tambo integrates with five major AI providers:
| Provider | Description | Best For |
| ------------- | ---------------------------------------------------------------------------------- | ------------------------------------------------------------------- |
| **OpenAI** | Industry-leading models including GPT-4.1, GPT-5, GPT-5.1, and o3 reasoning models | General-purpose tasks, reasoning, and state-of-the-art performance |
| **Anthropic** | Claude models with strong safety and reasoning capabilities | Complex reasoning, analysis, and safety-critical applications |
| **Google** | Gemini models with multimodal support and extended thinking capabilities | Multimodal tasks, vision-based applications, and advanced reasoning |
| **Groq** | High-speed inference with Llama models optimized for throughput | Low-latency applications and real-time processing |
| **Mistral** | Fast, efficient open-source models with strong performance | Cost-effective alternatives with reliable performance |
## Configuring Providers
All model providers are configured through your Tambo Cloud dashboard:
1. Navigate to **Dashboard** → **Project** → **Settings** → **LLM Providers**
2. Select your desired provider
3. Choose a model from the available options
4. (Optional) Add custom parameters for fine-tuned behavior
You can configure multiple providers in a single project and switch between
them as needed. This is useful for testing different models or optimizing for
different use cases.
## Model Status Labels
Each model carries a status label indicating how thoroughly it has been tested with Tambo:
* **Tested** - Validated on common Tambo tasks; recommended for production
* **Untested** - Available but not yet validated; use with caution and test in your context
* **Known Issues** - Usable but with observed behaviors worth noting
For detailed information about each label and specific model behaviors, see [Labels](/models/labels).
Streaming may behave inconsistently in models other than OpenAI. We're aware
of the issue and actively working on a fix. Please proceed with caution when
using streaming on non-OpenAI models.
## Advanced Configuration
### Custom LLM Parameters
Fine-tune model behavior with custom parameters like temperature, max tokens, and provider-specific settings. This allows you to optimize models for your specific use case—whether you need deterministic responses for analysis or creative outputs for generation.
**Common parameters across all providers:**
* `temperature` - Control randomness (0.0-1.0)
* `maxOutputTokens` - Limit response length
* `topP` - Nucleus sampling threshold
* `presencePenalty` - Encourage staying on topic
* `frequencyPenalty` - Reduce repetition
Learn more in [Custom LLM Parameters](/models/custom-llm-parameters).
### Reasoning Models
Advanced reasoning models from OpenAI (GPT-5, GPT-5.1, O3) and Google (Gemini 3.0 Pro, Gemini 3.0 Deep Think) expose their internal thinking process. These models excel at complex problem-solving by spending additional compute time analyzing problems before generating responses.
Configure reasoning capabilities through your project's LLM provider settings to enable:
* Multi-step problem decomposition
* Solution exploration and verification
* Detailed reasoning token access
* Adaptive thinking time (for supported models)
See [Reasoning Models](/models/reasoning-models) for detailed configuration guides.
## Quick Links
Understand model testing status and known behaviors
Fine-tune model behavior with temperature, tokens, and more
Configure advanced reasoning capabilities for complex tasks
## Next Steps
* **Getting Started**: Choose a provider and configure it in your project settings
* **Optimize Performance**: Use custom parameters to fine-tune responses for your use case
* **Explore Reasoning**: Enable reasoning on supported models for complex tasks
* **Monitor Usage**: Track model performance and costs in your dashboard
For comprehensive API and integration guidance, explore the [API Reference](/api-reference) and [Concepts](/concepts) sections.