The Knowledge Assistant supports multiple large language models (LLMs) from different providers. Each model has different characteristics in terms of speed, quality, and cost. You can select your preferred model from the settings menu.
Google's latest fast model optimized for speed while maintaining high quality. Excellent balance of performance and response time for most use cases.
Google's most capable reasoning model with enhanced analytical abilities. Better for complex questions requiring multi-step reasoning.
Anthropic's balanced model with strong reasoning and writing capabilities. Excellent for nuanced analysis and detailed explanations.
Anthropic's fastest model, optimized for quick responses while maintaining good quality. Ideal for simple queries and high-throughput applications.
| Model | Speed | Quality | Best For |
|---|---|---|---|
| Gemini 3 Flash | ★★★★★ | ★★★★☆ | Daily use, most queries |
| Gemini 3 Pro | ★★★☆☆ | ★★★★★ | Complex analysis, reasoning |
| Claude Sonnet 4.5 | ★★★☆☆ | ★★★★★ | Detailed writing, nuance |
| Claude Haiku 4.5 | ★★★★★ | ★★★☆☆ | Quick lookups, simple queries |
Different models have different costs per token. While the Knowledge Assistant manages this internally, understanding the relative costs can help you choose:
Knowledge+ mode uses Google Search grounding, which is only available through Google's Gemini models. If you switch to Knowledge+ mode while using a Claude model, the system will automatically switch to Gemini Pro (default). Gemini Flash is used only when you explicitly request a fast model.