Agentastic supports multiple AI models, each optimized for different tasks. Understanding which model to use and when will help you get the best performance and results.
Model Types
Agentastic categorizes models into six capability types, each optimized for different tasks:
Chat Models 💬
General-purpose conversational models for everyday tasks:
- Best for: Quick queries, text generation, casual conversations
- Speed: Fast response times
- Tools: Limited tool access
- Examples: GPT-4o, Claude Haiku, Agentastic Chat
When to use: Default choice for most interactions, simple text expansions, quick questions.
Vision Models 👁
Models that can understand and analyze images:
- Best for: Screenshot analysis, image descriptions, visual Q&A
- Speed: Moderate response times
- Tools: Standard tools plus image processing
- Examples: GPT-4o, Claude Sonnet, Qwen3 VL, Agentastic Vision
When to use: Analyzing screenshots, understanding diagrams, extracting text from images.
Agent Models ∞
Advanced models with reasoning and tool-use capabilities:
- Best for: Complex tasks, multi-step workflows, tool integration
- Speed: Slower but more capable
- Tools: Full access to all tools (calendar, email, search, files)
- Examples: Agentastic Agent, Claude Opus, GPT-5, Qwen3 Max
When to use: Complex research, calendar management, email composition, file operations.
Automation Models 🖱️
Specialized models for desktop control and automation:
- Best for: GUI automation, desktop control, system tasks
- Speed: Variable depending on task
- Tools: Desktop control, mouse/keyboard automation
- Examples: Agentastic Computer Use, Claude Sonnet, OpenAI Computer Use
When to use: Automating repetitive desktop tasks, UI testing, complex workflows.
Thinking Models 🧠
Models with advanced reasoning and "thinking" capabilities:
- Best for: Deep reasoning, complex problem-solving, multi-step analysis
- Speed: Slower due to extended reasoning
- Tools: Full tool access with enhanced reasoning
- Examples: OpenAI O3, DeepSeek R1, Qwen3 Thinking, Claude Opus
When to use: Complex problem-solving, research requiring deep reasoning, multi-step analysis.
Image Generation Models 🎨
Models that can generate and edit images:
- Best for: Creating images from text, image editing
- Speed: Variable depending on complexity
- Tools: Image generation and editing
- Examples: Agentastic Imagine, Gemini Banana, Hunyuan v3, SeeDream v4
When to use: Creating images from descriptions, editing existing images, visual content generation.
Available Models
Agentastic Recommended Models 💜
These are Agentastic's optimized models, recommended for most users:
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| Agentastic Agent 💜 | Agent, Chat | 256K | Complex workflows, tool-based tasks | /model:agent |
| Agentastic Chat 💜 | Chat | 128K | Quick questions, casual chat | /model:chat |
| Agentastic Auto 💜 | Chat | 128K | When you want the system to choose | /model:auto |
| Agentastic Vision 💜 | Chat, Agent, Vision | 200K | Screenshot analysis, image understanding | /model:vision |
| Agentastic Writer 💜 | Chat, Agent, Vision | 128K | Writing, editing, content creation | /model:writer |
| Agentastic Imagine 💜 | Image Generation | 128K | Creating images from text | /model:imagine |
| Agentastic Computer Use 💜 | Automation | 128K | GUI automation, desktop tasks | /model:computer |
| Gemini Nano-Banana 💜 | Image Generation, Vision, Chat | 128K | Image understanding and generation | /model:banana |
| GPT-5 Chat 💜 | Chat, Thinking | 128K | General-purpose tasks with reasoning | /model:gpt |
| OpenAI O3 💜 | Chat, Thinking | 200K | Deep reasoning, complex problem-solving | /model:o3 |
| Grok Code Fast 💜 | Chat, Agent, Thinking | 256K | Code-related tasks | /model:grok-code |
| Qwen3 Coder 💜 | Chat, Agent, Thinking | 128K | Code generation, debugging, review | /model:qwen3-coder |
| Qwen3 Max 💜 | Chat, Agent, Thinking | 256K | Complex reasoning tasks | /model:qwen3-max |
| Qwen3 Next 💜 | Chat, Agent, Thinking | 262K | General-purpose advanced tasks | /model:qwen3-next |
| Qwen3 VL 💜 | Chat, Agent, Vision, Automation, Thinking | 131K | Tasks requiring vision and reasoning | /model:qwen3-vl |
OpenAI Models
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| GPT-5 | Chat, Agent, Vision, Thinking | 272K | Complex tasks requiring multiple capabilities | /model:gpt-5 |
| GPT-5 Chat - Mini | Chat, Agent, Vision | 272K | Quick tasks with GPT-5 capabilities | /model:gpt-5-mini |
| GPT-4o | Chat, Agent, Vision | 128K | General-purpose tasks, quick responses | /model:gpt-4o |
| GPT-4o-mini | Chat, Agent, Vision, Thinking | 128K | Quick tasks with good quality | /model:gpt-4o-mini |
| OpenAI O3 | Chat, Thinking | 200K | Deep reasoning, complex problem-solving | /model:o3 |
| OpenAI Computer Use | Automation | 8K | GUI automation, desktop tasks | /model:openai-computer |
| OpenAI gpt-oss-120b | Chat, Agent, Thinking | 128K | General tasks with open-source model | /model:gpt-oss-120b |
| OpenAI gpt-oss-20b | Chat, Agent, Thinking | 128K | Quick tasks with open-source model | /model:gpt-oss-20b |
Anthropic (Claude) Models
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| Claude Opus | Chat, Vision, Thinking | 200K | Research, advanced workflows, deep reasoning | /model:opus |
| Claude Sonnet 4.5 | Chat, Agent, Vision, Automation, Thinking | 200K | Most tasks, image analysis, automation | /model:sonnet |
| Claude Haiku 4.5 | Chat, Vision, Agent, Automation, Thinking | 200K | Simple queries, quick tasks | /model:haiku |
Google Models
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| Google Gemini 3 Pro (Preview) | Chat, Agent, Vision, Thinking | 1M in / 64K out | Deep reasoning, Deep Think mode, Jan 2025 knowledge | /model:gemini |
| Google Gemini 2.5 Flash | Chat, Agent, Vision | 1M | Quick tasks, segmentation, lower-latency 1M context | /model:flash |
| Google Gemma 3 27B | Chat, Vision | 128K | General tasks with vision | /model:gemma |
Gemini 3 update (Nov 18, 2025)
- Launch blog highlights that Gemini 3 Pro now tops major benchmarks (1501 Elo on LMArena, 81% on MMMU-Pro, 87.6% on Video-MMMU, 72.1% on SimpleQA Verified) and introduces a Deep Think mode that reaches 41% on Humanity's Last Exam (no tools) and 45.1% on ARC-AGI-2 with code execution. Deep Think is rolling out to Gemini Ultra subscribers after safety testing.
- The Gemini 3 API guide documents the
gemini-3-pro-previewmodel ID, a 1M input / 64K output window, a Jan 2025 knowledge cutoff, and pricing of $2 / $12 per 1M tokens below 200K tokens ($4 / $18 after). It also introduces thethinking_levelparameter (lowfor lower latency/cost,highdefault for Deep Think depth—mediumis coming soon) and the newmedia_resolution_{low|medium|high}knobs: images default to 1120 tokens athigh, PDFs saturate at 560 tokens withmedium, and video frames cost 70 tokens atlow/medium(280 tokens if you forcehighfor dense OCR).- Gemini 3 Pro does not yet support pixel-level segmentation; keep
/model:flash(Gemini 2.5 Flash with thinking disabled) or Gemini Robotics-ER 1.5 for those workloads per Google's guidance.
DeepSeek Models
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| DeepSeek v3.2 | Chat, Agent, Thinking | 131K | General tasks with reasoning | /model:deepseek |
| DeepSeek v3 R1 0528 | Chat, Thinking | 131K | Deep reasoning tasks | /model:deepseek-r1 |
Grok Models (x-ai)
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| x-ai Grok 4 | Chat, Agent, Thinking | 256K | Complex tasks with reasoning | /model:grok |
| x-ai Grok 4 Fast | Chat, Agent | 2M | Quick tasks with very long context | /model:grok-fast |
| x-ai Grok Code Fast 💜 | Chat, Agent, Thinking | 256K | Code-related tasks | /model:grok-code |
Qwen Models
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| Qwen3 | Agent, Chat, Thinking | 128K | General tasks with reasoning | /model:qwen3 |
| Qwen3 Thinking | Chat, Agent, Thinking | 262K | Deep reasoning tasks | /model:qwen3-thinking |
| Qwen3 Max 💜 | Chat, Agent, Thinking | 256K | Complex reasoning tasks | /model:qwen3-max |
| Qwen3 Next 💜 | Chat, Agent, Thinking | 262K | Advanced tasks | /model:qwen3-next |
| Qwen3 Next Thinking 💜 | Chat, Agent, Thinking | 262K | Deep reasoning with latest model | /model:qwen3-next-thinking |
| Qwen3 Coder 💜 | Chat, Agent, Thinking | 128K | Code generation and analysis | /model:qwen3-coder |
| Qwen3 VL 💜 | Chat, Agent, Vision, Automation, Thinking | 131K | Vision and reasoning tasks | /model:qwen3-vl |
| Qwen3 VL Thinking 💜 | Chat, Agent, Vision, Automation, Thinking | 131K | Vision tasks requiring deep reasoning | /model:qwen3-vl-thinking |
Meta (Llama) Models
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| Meta LLama4 Maverick | Chat, Vision, Thinking | 128K | General tasks with vision and reasoning | /model:llama4-maverick |
| Meta LLama4 Scout | Chat, Vision, Thinking | 128K | Quick tasks with vision | /model:llama4-scout |
Other Providers
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| Mistral Nemo | Chat, Agent, Vision | 131K | General tasks | /model:nemo |
| MoonshotAI Kimi-K2 | Chat, Agent | 256K | Quick tasks | /model:kimi |
| Z.ai GLM 4.6 | Chat, Agent | 202K | Long documents | /model:glm |
| Z.ai GLM Air 4.5 | Chat, Agent | 128K | Quick tasks | /model:glm-air |
| Z.ai GLM 4.5 Vision | Chat, Vision | 65K | Image analysis | /model:glm-vision |
| Perplexity Sonar | Chat, Thinking | 128K | Research and reasoning (use instead of /model:sonar-reasoning) | /model:sonar |
| Perplexity Sonar Pro | Chat, Thinking | 128K | Advanced research and reasoning | /model:sonar-pro |
| Minimax M2 | Chat, Vision, Agent | 196K | General tasks with vision | /model:minimax |
Image Generation Models
| Model | Capabilities | Context | Best For | Command |
|---|---|---|---|---|
| Agentastic Imagine 💜 | Image Generation | 128K | Creating and editing images | /model:imagine |
| Gemini Nano-Banana 💜 | Image Generation, Vision, Chat | 128K | Image understanding and generation | /model:banana |
| Tencent Hunyuan v3 | Image Generation | 128K | Creating images from text | /model:hunyuan |
| ByteDance SeeDream v4 | Image Generation | 128K | Creating images | /model:seedream |
Model Selection
Quick Selection in Launcher
Use the /model: tag to quickly switch models:
- Type
/model:in the launcher - Browse available models with arrow keys
- Select with Enter or click
- Continue with your prompt
Example:
/model:opus Analyze this research paper and summarize key findings
/model:agent What's on my calendar today?
/model:vision What do you see in this screenshot?
Setting Default Models
Configure your preferred models in Settings:
- Open Settings (
⌘,) - Go to AI Models tab
- Set defaults for each model type:
- Default Chat Model
- Default Vision Model
- Default Agent Model
- Default Automation Model
- Default Thinking Model
- Default Image Generation Model
Model-Specific Tags
Some tags automatically suggest appropriate models:
/vision- Activates vision-capable models/agent- Enables agent mode with tool access/automation- Uses automation-capable models/imagine- Uses image generation models
Configuration
Models Configuration File
Advanced users can customize models in ~/Library/Application Support/agentastic/config/models.yml:
active_models:
- value: agentastic:agent
label: Agentastic Agent 💜
description: Designed for advanced agentic tasks
context_length: 256000
capabilities:
- agent
- chat
fallback_model: agentastic:sonnet
- value: agentastic:chat
label: Agentastic Chat 💜
description: The friendly neighborhood model
context_length: 128000
capabilities:
- chat
fallback_model: agentastic:sonnet
- value: agentastic:vision
label: Agentastic Vision 💜
description: Optimized vision analysis model
context_length: 200000
capabilities:
- chat
- agent
- vision
fallback_model: agentastic:sonnet
API Keys
Configure API keys for different providers:
- Open Settings → AI Models
- Click Configure API Keys
- Enter keys for:
- OpenAI (for GPT models)
- Anthropic (for Claude models)
- Google (for Gemini models)
- DeepSeek, Grok, Qwen, and other providers (if configured)
Security Note: API keys are stored securely in your macOS keychain.
Model Selection Strategy
For Different Tasks
Writing Emails
- Quick email: Agentastic Chat, GPT-4o, Claude Haiku
- Important email: Agentastic Writer, Claude Sonnet, GPT-5
- Complex negotiation: Claude Opus, Qwen3 Max
Research & Analysis
- Quick fact check: Agentastic Auto, GPT-4o
- Deep research: Claude Opus, Qwen3 Max with
/agent - Visual analysis: Agentastic Vision, GPT-4o, Qwen3 VL
Code & Technical
- Quick snippets: Agentastic Chat, GPT-4o
- Complex debugging: Qwen3 Coder, Grok Code Fast, Claude Opus
- Code review: Claude Sonnet, Agentastic Writer
Automation
- Simple tasks: Agent models with tools
- Desktop control: Agentastic Computer Use, OpenAI Computer Use
- Multi-step workflows: Agentastic Agent, Claude Opus with
/agent
Performance vs. Quality
Balance speed and capability based on your needs:
Fast Response Needed → Agentastic Chat, GPT-4o, Claude Haiku, GPT-4o-mini
Balanced Performance → Agentastic Auto, Claude Sonnet, GPT-5, Gemini 3 Pro (Preview)
Best Quality → Claude Opus, GPT-5, Qwen3 Max, OpenAI O3
Special Capabilities → Vision models for images (Agentastic Vision, Qwen3 VL), Automation models for desktop control (Agentastic Computer Use), Thinking models for deep reasoning (O3, DeepSeek R1), Image generation (Agentastic Imagine, Gemini Banana)
Cost Considerations
Different models have different pricing:
Cost Tiers (Relative)
- Low Cost: Agentastic Chat, Claude Haiku, GPT-4o-mini (for quick tasks)
- Medium Cost: Agentastic Auto, Claude Sonnet, GPT-4o, Gemini 2.5 Flash
- Higher Cost: Claude Opus, GPT-5, Gemini 3 Pro (Preview), Qwen3 Max, Agentastic Agent
- Specialized: Automation models, Thinking models, Image generation models (variable)
Optimizing Costs
- Use appropriate models: Don't use Opus for simple tasks
- Leverage chat mode: Agent mode uses more tokens
- Clear context: Start new conversations when context isn't needed
- Local models: Consider Ollama for high-volume use
Model Limitations
Context Windows
Different models have different context limits:
- Small (8K-65K): OpenAI Computer Use (8K), GLM Vision (65K)
- Standard (128K-131K): Most models including GPT-5, GPT-4o, DeepSeek
- Large (200K-256K): Claude models (200K), Agentastic Agent (256K), Grok models (256K), Qwen3 Max (256K)
- Very Large (262K): Qwen3 Thinking, Qwen3 Next models
- Massive (1M+): Google Gemini 3 Pro (Preview) / Gemini 2.5 Flash (1M tokens), Grok 4 Fast (2M tokens)
Rate Limits
API providers impose rate limits:
- Per minute: Number of requests
- Per day: Total token usage
- Concurrent: Parallel requests
Agentastic automatically manages rate limits and queues requests when necessary.
Feature Support
Not all models support all features:
| Feature | Chat | Vision | Agent | Automation | Thinking | Imagine |
|---|---|---|---|---|---|---|
| Text generation | ✅ | ✅ | ✅ | ✅ | ✅ | ✅ |
| Image understanding | ❌ | ✅ | ✅* | ✅* | ✅* | ✅ |
| Tool use | Limited | Limited | ✅ | ✅ | ✅ | ❌ |
| Desktop control | ❌ | ❌ | ❌ | ✅ | ❌ | ❌ |
| Multi-step reasoning | Limited | Limited | ✅ | ✅ | ✅ | ❌ |
| Advanced reasoning | ❌ | ❌ | Limited | Limited | ✅ | ❌ |
| Image generation | ❌ | ❌ | ❌ | ❌ | ❌ | ✅ |
*Depends on specific model
Troubleshooting
Model Not Available
If a model isn't showing up:
- Check API key is configured
- Verify internet connection
- Check provider service status
- Ensure model is enabled in settings
Poor Results
If a model isn't performing well:
- Try a more capable model
- Use appropriate tags (
/agent,/vision) - Provide clearer prompts
- Add relevant context via memory channels
Rate Limit Errors
If you hit rate limits:
- Wait for the reset period (usually 1 minute)
- Switch to a different model
- Reduce request frequency
- Consider upgrading API tier
Best Practices
- Start with Agentastic Auto or recommended models (💜) for most tasks
- Use vision models only when analyzing images
- Enable agent mode when you need tools
- Reserve automation models for desktop tasks
- Use thinking models for complex reasoning problems
- Use image generation models for creating visual content
- Test different models to find what works best for your workflow
- Monitor usage to optimize costs
What's Next?
Now that you understand models:
- Core Features - Explore all capabilities
- AI Commands - Master special tags
- Advanced Settings - Fine-tune configuration
- Troubleshooting - Solve common issues
Pro Tip: Use /model: frequently to experiment with different models. You'll quickly learn which models work best for your specific use cases.