AI Models
TeXRA supports a variety of language models from different providers, allowing you to choose the best fit for your task's complexity, required speed, and budget (think of it as choosing your research assistant's brain!). This guide provides an overview of the models available by default.
Model Providers Overview
TeXRA primarily integrates with models from:
- Anthropic (Claude family)
- OpenAI (GPT and O-series families)
- Google (Gemini family)
- Other Providers (via OpenRouter, including Grok, DeepSeek)
You can select the desired model from the dropdown list in the TeXRA UI.
Default Model Selection
Here's a quick comparison of the models available by default in TeXRA:
Anthropic Models
Known for strong instruction following and context handling.
Model ID | Key Strength / Use Case | Relative Cost | Relative Speed | Notes |
---|---|---|---|---|
opus | Highest quality, complex tasks | $$$$ | Slow | |
sonnet37 | Strong all-rounder, good context | $$$ | Medium | |
sonnet37T | sonnet37 with explicit reasoning steps | $$$ | Medium | Good for math, complex logic |
sonnet35 | Good balance of quality/cost (older Sonnet) | $$$ | Medium |
OpenAI Models
Known for strong reasoning and creative capabilities.
Model ID | Key Strength / Use Case | Relative Cost | Relative Speed | Notes |
---|---|---|---|---|
o1 | Advanced reasoning, math, figures | $$$$ | Slow | Explicit reasoning |
gpt45 | High quality, vision (Preview) | $$$$ | Medium | |
gpt41 | Long-context vision, powerful | $$$ | Medium | 1M tokens context |
gpt41- | Long-context vision, cost-effective | $$ | Medium | 1M tokens context, mini |
gpt41-- | Long-context vision, cheapest | $ | Medium | 1M tokens context, nano |
gpt4o | Strong all-rounder, vision | $$$ | Medium | Good default choice |
gpt4ol | Latest gpt4o , potentially better | $$$ | Medium | |
o3- | Fast reasoning | $$$ | Fast | o3-mini |
o1- | Fast reasoning (smaller o1 ) | $$$ | Fast | o1-mini |
Google Models
Known for large context windows, multimodality, and speed/cost efficiency.
Model ID | Key Strength / Use Case | Relative Cost | Relative Speed | Notes |
---|---|---|---|---|
gemini25p | Strong reasoning, vision, large context | $$$ | Medium | Latest Pro model |
gemini2p | Good reasoning, vision, very large context | $$ | Medium | |
gemini25f | Fast reasoning, large context | $$ | Fast | Latest Flash model |
gemini2fT | gemini2f with explicit reasoning steps | $ | Fast | |
gemini2f | Fastest, most cost-effective, vision | $ | Very Fast | Good for simple tasks, native PDF/audio |
DeepSeek Models
Strong technical and coding performance, cost-effective.
Model ID | Key Strength / Use Case | Relative Cost | Relative Speed | Notes |
---|---|---|---|---|
DSV3 | Good coding & general tasks | $ | Fast | DeepSeek V3 Chat |
DSR1 | Advanced reasoning | $$ | Medium | DeepSeek R1 |
Grok / xAI Models
Large context models from xAI.
Model ID | Key Strength / Use Case | Relative Cost | Relative Speed | Notes |
---|---|---|---|---|
grok3 | Large context, alternative reasoning | $$$ | Medium | xAI Grok 3 |
grok3- | Faster Grok 3 (mini) | $$ | Fast | xAI Grok 3 Mini |
Other Models (Available Primarily via OpenRouter)
These models are generally accessed by enabling OpenRouter in settings.
Model ID | Key Strength / Use Case | Provider | Relative Cost | Relative Speed |
---|---|---|---|---|
llama31 | Strong open model, large context | Meta | $$$ | Medium |
qvq-72b | Strong multi-lingual | Qwen/Alibaba | $$ | Medium |
Relative Cost/Speed are estimates: $ = Low/Fast, $$$$ = High/Slow.
Choosing the Right Model
Consider these factors:
- Task Complexity: Simple corrections might only need a
$
/Fast model (gemini2f
), while complex paper transformations benefit from$$$$
/Slow models (opus
,o1
). - Budget: Use cost indicators ($ - $$$$) to guide selection.
- Speed: If quick turnaround is needed, prefer Fast/Very Fast models.
- Special Capabilities: Do you need explicit reasoning (
sonnet37T
,gemini2fT
,o1
,o3-
,o1-
,DSR1
), vision (gpt4o
,gemini*
), native PDF/audio (gemini*
), or very large context (gemini*
,gpt41
)?
Experimentation is often key to finding the best model for your specific needs and writing style.
Model Configuration
You can customize which models appear in the TeXRA dropdown list via VS Code Settings (Ctrl+,
). Search for texra.models
and edit the JSON array. Here are the defaults:
Model Availability
The specific models available by default and their identifiers (sonnet37
, gpt4o
, etc.) are maintained by the TeXRA developers and may change in future updates based on new releases and performance evaluations.
"texra.models": [
"sonnet37T",
"sonnet37",
"sonnet35",
"opus",
"o3-",
"o1",
"o1-",
"gpt45",
"gpt4o",
"gpt4ol",
"gemini25p",
"gemini25f",
"gemini2p",
"gemini2f",
"gemini2fT",
"DSV3",
"DSR1",
"grok3",
"grok3-"
]
Using OpenRouter
To access models not directly integrated (like Llama or Qwen), find alternative pricing, or ensure access if a direct API key isn't available, you can use OpenRouter.
- Get an OpenRouter API key.
- Add the key using the
TeXRA: Set API Key
command (select OpenRouter). - Enable OpenRouter in VS Code Settings:
"texra.model.useOpenRouter": true
.
When enabled, TeXRA will route API calls for all compatible models (including Anthropic, OpenAI, Google, DeepSeek, Grok, etc., if supported by OpenRouter) through OpenRouter instead of their direct APIs.
Streaming Support
For long responses or reasoning-heavy models, you can enable streaming to see incremental results. This is often more robust for complex tasks.
Configure streaming in VS Code Settings:
// General streaming toggle (applies if specific model type toggle isn't set)
"texra.model.useStreaming": false,
// Specific toggle for Anthropic reasoning models
"texra.model.useStreamingAnthropicReasoning": false,
// Specific toggle for OpenAI reasoning models
"texra.model.useStreamingOpenAIReasoning": false
// Similar configuration for Google/DeepSeek/OpenRouter models
Next Steps
- Built-in Agents: See which agents work well with different models.
- Configuration: Learn about other model-related settings like streaming.