# Labels URL: /models/labels Streaming may behave inconsistently in models other than **OpenAI**. We're aware of the issue and are actively working on a fix. Please proceed with caution when using streaming on non-OpenAI models. Models in tambo carry a **status label**, shown when you select a model from the LLM settings\ (**Dashboard → Project → Settings → LLM Providers**). ### Why Use Labels? * **Set expectations**: Understand tambo’s confidence level for each model. * **Guide selection**: Prefer `tested` models for production; approach others with care. * **Highlight caveats**: `known-issues` labels call out specific behaviors we've observed. ### Label Definitions | Label | Meaning | | -------------- | ------------------------------------------------------------------ | | `tested` | Validated on common tambo tasks. Recommended for most workflows. | | `untested` | Available, but not yet validated. Use it—but test in your context. | | `known-issues` | Usable, but we’ve observed behaviors worth noting (see below). | ### Observed Behaviors & Notes These behaviors were noted during testing on common tambo tasks. The models below are still usable—just keep the caveats in mind. #### Google * **Gemini 2.5 Pro / 2.5 Flash / 2.0 Flash / 2.0 Flash Lite**: * May occasionally resist rendering as requested. Sometimes it completes the request, but behavior can be inconsistent. * Try clarifying instructions (e.g., “Return a bulleted list only”). * Outputs may have formatting quirks. Be cautious when structure matters. #### Anthropic * **Claude 3.5 Haiku**: * May fail to fully render components, even when receiving the correct data. * *Example:* When rendering a graph component, it may leave it in a loading state without streaming the data into props. #### Mistral * **Mistral Large 2.1 / Medium 3**: * Similar to Gemini, may inconsistently follow rendering instructions. * Try clarifying the prompt structure. * Formatting idiosyncrasies can occur—validate outputs where structure is important. For production-critical formatting, use **Tested** models and validate outputs. When using **Untested** or **Known Issues** models, run a small prompt suite to check behavior in your specific workload. ### Usage Patterns * **Prefer `tested`** models for reliability. If using others, test with your use case. * **Use inline notes** in the picker to spot caveats quickly. ### Integration You can change providers and models at the project level under **LLM Provider Settings**. tambo will apply your token limits and defaults accordingly.