Available models

To add a model to the Chat and ⌘K selection menu, enable it from Cursor Settings > Models.

ModelProviderPremiumAgentPrice6
claude-3.7-sonnetAnthropic$0.04
claude-3.7-sonnet MAX 1Anthropic$0.05
claude-3.5-sonnetAnthropic$0.04
claude-3.5-haiku 2Anthropic$0.01
claude-3-opus 3Anthropic$0.10
cursor-smallCursorFree
deepseek-v3FireworksFree
deepseek-r1Fireworks$0.04
gemini-2.5-pro-expGoogle$0.04
gemini-2.0-pro-expGoogle$0.04
gpt-4oOpenAI$0.04
gpt-4o-mini 4OpenAIFree
gpt-4.5-previewOpenAI$2.00
o1OpenAI$0.40
o1-mini 3OpenAI$0.10
o3-mini 2, 5OpenAI$0.01
grok-2xAI$0.04

1 Tool calls charged like requests
2 1/3 request
3 10 requests/day included with paid plan

4 500 requests/day with free plan
5 High reasoning effort
6 Prices are higher for long context window requests

Premium models

Premium models can be used with the 500 requests included with your monthly Pro or Business subscription.

Once you’ve exhausted your 500 monthly requests, Cursor will continue to serve you premium model requests, but may delay response time and/or limit access to some models when the platform is under high load.

To avoid delays and limited access, you can enable usage-based pricing for Premium models from Settings and pay per request after exhausting your monthly allotment of 500.

Agentic models

Agentic models can be used with Chat’s Agent mode. These models are highly capable at making tool calls and perform best with Agent.

Submitting an Agent prompt with up to 25 tool calls consumes one request. If your request extends beyond 25 tool calls, Cursor will ask if you’d like to continue which will consume a second request.

Non-premium models

Those models not designated Premium are pay-as-you-go and can be used by enabling usage-based pricing from Settings. The 500 montly requests cannot be used for these models.

MAX

Models offered in MAX mode have enhanced capabilities with larger context windows and expanded reasoning.

Currently offered as an option for Claude 3.7 Sonnet, MAX mode provides a 200k token context window, 200 Agent tool call limit, and the ability to process up to 750 lines per file read operation.

When operated as an Agent, each tool call in MAX mode is charged as a separate request in addition to the initial prompt request.

Selecting a model

The input interface in Chat and ⌘K includes a model selection menu which lists the models enabled from Settings > Models.

Auto-select

Enabling Auto-select configures Cursor to select the premium model best fit for the immediate task and with the highest reliability based on current demand. This feature can detect degraded output performance and automatically switch models to resolve it.

Thinking

Enabling Thinking limits the list of models to reasoning models which think through problems step-by-step and have deeper capacity to examine their own reasoning and correct errors.

These models often perform better on complex reasoning tasks, though they may require more time to generate their responses.

Context windows

A context window is the maximum span of text and code an LLM can consider at once, including both the input prompt and output generated by the model.

Each Chat session in Cursor maintains its own context window. The more prompts, attached files, and responses included in a session, the larger the context window grows.

Cursor actively optimizes the context window as the Chat session progresses, intelligently pruning non-essential content while preserving critical code and conversation elements.

For best results, it’s recommended you take a purpose-based approach to Chat session management, starting a new session for each unique task.

Large context and pricing

Large context mode allows Cursor to process more text and code per session, handling larger files and complex tasks while maintaining critical context.

Using large context doubles the request price compared to standard context sessions. This pricing reflects the increased computational resources required to process and analyze larger amounts of information.

Large context can be enabled in two ways:

  • Manually by checking ‘Large context’ from Settings > Features
  • Automatically when a Chat session grows long and/or includes large file attachments; in this case, large context charges will be indicated by a ⚠︎ icon appearing in the input interface

You can monitor per requests costs in real time from the Settings page.

Context window sizes

ModeContext Window
Agent: claude-3.7-sonnet MAX200,000 tokens
Agent: claude-3.7-sonnet120,000 tokens
Agent: Other Models60,000 tokens
⌘K10,000 tokens

These thresholds are subject to change as Cursor further optimizes its context capabilities.

Model hosting

Models are hosted on US-based infrastructure by the model’s provider, a trusted partner or Cursor.

When Privacy Mode is enabled from Settings, Cursor nor the model providers will store your data with all data deleted after each request is processed. For further details see our Privacy, Privacy Policy, and Security pages.

Was this page helpful?