Skip to main content
Access: Open an agent and go to GeneralThinking.

What the Model Does

The AI model reads the conversation transcript and decides what to say next. It follows your prompt, retrieves from knowledge bases, and triggers tools like transfers and bookings. Choosing the right model means balancing response quality, latency, and cost.
Some models incur additional per-minute charges on top of the base rate. Check the cost indicator next to each model in the catalog, or see Premium Features for details.

PresetWhat it does
IntelligentHigher-quality responses with more latency. Use for complex reasoning, multi-step conversations, or brand-sensitive interactions.
BalancedGood quality and speed for most use cases. Recommended for most agents.
FastLowest latency. Use for high-volume or simple routing and qualification flows.
Switch only if testing shows you need more quality or speed.

Response Style

The Response Style slider appears under General → Thinking when the selected model supports adjustable temperature. Temperature controls how consistent or varied the agent’s responses are:
RangeBehaviorUse for
0.0Fully deterministicMost agents — maximizes reliability for tool calling
0.1–0.3Slight variationAgents that need natural phrasing variation
0.4–0.7More creativePersonality-driven agents where consistency matters less
0.8+UnpredictableAvoid in production
Use 0.0 for agents that transfer calls, book appointments, or call APIs. Higher temperature reduces tool execution reliability.

Next Steps

Select Voice

Choose how your agent sounds to callers

Transcriber

Configure the speech-to-text layer

Voice Settings

Fine-tune speed, stability, and style

Test Your Agent

Test model performance with web calls