Skip to content

AI Models

TeXRA supports a variety of language models from different providers, allowing you to choose the best fit for your task's complexity, required speed, and budget (think of it as choosing your research assistant's brain!). This guide provides an overview of the models available by default.

Model Providers Overview

TeXRA primarily integrates with models from:

  1. Anthropic (Claude family)
  2. OpenAI (GPT and O-series families)
  3. Google (Gemini family)
  4. Other Providers (via OpenRouter, including Grok, DeepSeek)

You can select the desired model from the dropdown list in the TeXRA UI.

Default Model Selection

Here's a quick comparison of the models available by default in TeXRA:

Anthropic Models

Known for strong instruction following and context handling.

Model IDKey Strength / Use CaseRelative CostRelative SpeedNotes
opusHighest quality, complex tasks$$$$Slow
sonnet37Strong all-rounder, good context$$$Medium
sonnet37Tsonnet37 with explicit reasoning steps$$$MediumGood for math, complex logic
sonnet35Good balance of quality/cost (older Sonnet)$$$Medium

OpenAI Models

Known for strong reasoning and creative capabilities.

Model IDKey Strength / Use CaseRelative CostRelative SpeedNotes
o1Advanced reasoning, math, figures$$$$SlowExplicit reasoning
gpt45High quality, vision (Preview)$$$$Medium
gpt41Long-context vision, powerful$$$Medium1M tokens context
gpt41-Long-context vision, cost-effective$$Medium1M tokens context, mini
gpt41--Long-context vision, cheapest$Medium1M tokens context, nano
gpt4oStrong all-rounder, vision$$$MediumGood default choice
gpt4olLatest gpt4o, potentially better$$$Medium
o3-Fast reasoning$$$Fasto3-mini
o1-Fast reasoning (smaller o1)$$$Fasto1-mini

Google Models

Known for large context windows, multimodality, and speed/cost efficiency.

Model IDKey Strength / Use CaseRelative CostRelative SpeedNotes
gemini25pStrong reasoning, vision, large context$$$MediumLatest Pro model
gemini2pGood reasoning, vision, very large context$$Medium
gemini25fFast reasoning, large context$$FastLatest Flash model
gemini2fTgemini2f with explicit reasoning steps$Fast
gemini2fFastest, most cost-effective, vision$Very FastGood for simple tasks, native PDF/audio

DeepSeek Models

Strong technical and coding performance, cost-effective.

Model IDKey Strength / Use CaseRelative CostRelative SpeedNotes
DSV3Good coding & general tasks$FastDeepSeek V3 Chat
DSR1Advanced reasoning$$MediumDeepSeek R1

Grok / xAI Models

Large context models from xAI.

Model IDKey Strength / Use CaseRelative CostRelative SpeedNotes
grok3Large context, alternative reasoning$$$MediumxAI Grok 3
grok3-Faster Grok 3 (mini)$$FastxAI Grok 3 Mini

Other Models (Available Primarily via OpenRouter)

These models are generally accessed by enabling OpenRouter in settings.

Model IDKey Strength / Use CaseProviderRelative CostRelative Speed
llama31Strong open model, large contextMeta$$$Medium
qvq-72bStrong multi-lingualQwen/Alibaba$$Medium

Relative Cost/Speed are estimates: $ = Low/Fast, $$$$ = High/Slow.

Choosing the Right Model

Consider these factors:

  • Task Complexity: Simple corrections might only need a $/Fast model (gemini2f), while complex paper transformations benefit from $$$$/Slow models (opus, o1).
  • Budget: Use cost indicators ($ - $$$$) to guide selection.
  • Speed: If quick turnaround is needed, prefer Fast/Very Fast models.
  • Special Capabilities: Do you need explicit reasoning (sonnet37T, gemini2fT, o1, o3-, o1-, DSR1), vision (gpt4o, gemini*), native PDF/audio (gemini*), or very large context (gemini*, gpt41)?

Experimentation is often key to finding the best model for your specific needs and writing style.

Model Configuration

You can customize which models appear in the TeXRA dropdown list via VS Code Settings (Ctrl+,). Search for texra.models and edit the JSON array. Here are the defaults:

Model Availability

The specific models available by default and their identifiers (sonnet37, gpt4o, etc.) are maintained by the TeXRA developers and may change in future updates based on new releases and performance evaluations.

json
"texra.models": [
  "sonnet37T",
  "sonnet37",
  "sonnet35",
  "opus",
  "o3-",
  "o1",
  "o1-",
  "gpt45",
  "gpt4o",
  "gpt4ol",
  "gemini25p",
  "gemini25f",
  "gemini2p",
  "gemini2f",
  "gemini2fT",
  "DSV3",
  "DSR1",
  "grok3",
  "grok3-"
]

Using OpenRouter

To access models not directly integrated (like Llama or Qwen), find alternative pricing, or ensure access if a direct API key isn't available, you can use OpenRouter.

  1. Get an OpenRouter API key.
  2. Add the key using the TeXRA: Set API Key command (select OpenRouter).
  3. Enable OpenRouter in VS Code Settings: "texra.model.useOpenRouter": true.

When enabled, TeXRA will route API calls for all compatible models (including Anthropic, OpenAI, Google, DeepSeek, Grok, etc., if supported by OpenRouter) through OpenRouter instead of their direct APIs.

Streaming Support

For long responses or reasoning-heavy models, you can enable streaming to see incremental results. This is often more robust for complex tasks.

Configure streaming in VS Code Settings:

json
// General streaming toggle (applies if specific model type toggle isn't set)
"texra.model.useStreaming": false,

// Specific toggle for Anthropic reasoning models
"texra.model.useStreamingAnthropicReasoning": false,

// Specific toggle for OpenAI reasoning models
"texra.model.useStreamingOpenAIReasoning": false

// Similar configuration for Google/DeepSeek/OpenRouter models

Next Steps