Models
Cursor offers a variety of frontier models for use with Chat and ⌘K different pricing tiers
Available models
To add a model to the Chat and ⌘K selection menu, enable it from Cursor Settings > Models.
Model | Provider | Premium | Agent | Price6 |
---|---|---|---|---|
claude-3.7-sonnet | Anthropic | ✓ | ✓ | $0.04 |
claude-3.7-sonnet MAX 1 | Anthropic | ✓ | $0.05 | |
claude-3.5-sonnet | Anthropic | ✓ | ✓ | $0.04 |
claude-3.5-haiku 2 | Anthropic | ✓ | $0.01 | |
claude-3-opus 3 | Anthropic | ✓ | $0.10 | |
cursor-small | Cursor | Free | ||
deepseek-v3 | Fireworks | Free | ||
deepseek-r1 | Fireworks | ✓ | $0.04 | |
gemini-2.5-pro-exp | ✓ | ✓ | $0.04 | |
gemini-2.0-pro-exp | ✓ | $0.04 | ||
gpt-4o | OpenAI | ✓ | ✓ | $0.04 |
gpt-4o-mini 4 | OpenAI | ✓ | Free | |
gpt-4.5-preview | OpenAI | $2.00 | ||
o1 | OpenAI | $0.40 | ||
o1-mini 3 | OpenAI | $0.10 | ||
o3-mini 2, 5 | OpenAI | ✓ | ✓ | $0.01 |
grok-2 | xAI | ✓ | $0.04 |
1 Tool calls charged like requests
2 1/3 request
3 10 requests/day included with paid plan
4 500 requests/day with free plan
5 High reasoning effort
6 Prices are higher for long context window requests
Premium models
Premium models can be used with the 500 requests included with your monthly Pro or Business subscription.
Once you’ve exhausted your 500 monthly requests, Cursor will continue to serve you premium model requests, but may delay response time and/or limit access to some models when the platform is under high load.
To avoid delays and limited access, you can enable usage-based pricing for Premium models from Settings and pay per request after exhausting your monthly allotment of 500.
Agentic models
Agentic models can be used with Chat’s Agent mode. These models are highly capable at making tool calls and perform best with Agent.
Submitting an Agent prompt with up to 25 tool calls consumes one request. If your request extends beyond 25 tool calls, Cursor will ask if you’d like to continue which will consume a second request.
Non-premium models
Those models not designated Premium are pay-as-you-go and can be used by enabling usage-based pricing from Settings. The 500 montly requests cannot be used for these models.
MAX
Models offered in MAX mode have enhanced capabilities with larger context windows and expanded reasoning.
Currently offered as an option for Claude 3.7 Sonnet, MAX mode provides a 200k token context window, 200 Agent tool call limit, and the ability to process up to 750 lines per file read operation.
When operated as an Agent, each tool call in MAX mode is charged as a separate request in addition to the initial prompt request.
Selecting a model
The input interface in Chat and ⌘K includes a model selection menu which lists the models enabled from Settings > Models.
Auto-select
Enabling Auto-select configures Cursor to select the premium model best fit for the immediate task and with the highest reliability based on current demand. This feature can detect degraded output performance and automatically switch models to resolve it.
Thinking
Enabling Thinking limits the list of models to reasoning models which think through problems step-by-step and have deeper capacity to examine their own reasoning and correct errors.
These models often perform better on complex reasoning tasks, though they may require more time to generate their responses.
Context windows
A context window is the maximum span of text and code an LLM can consider at once, including both the input prompt and output generated by the model.
Each Chat session in Cursor maintains its own context window. The more prompts, attached files, and responses included in a session, the larger the context window grows.
Cursor actively optimizes the context window as the Chat session progresses, intelligently pruning non-essential content while preserving critical code and conversation elements.
For best results, it’s recommended you take a purpose-based approach to Chat session management, starting a new session for each unique task.
Large context and pricing
Large context mode allows Cursor to process more text and code per session, handling larger files and complex tasks while maintaining critical context.
Using large context doubles the request price compared to standard context sessions. This pricing reflects the increased computational resources required to process and analyze larger amounts of information.
Large context can be enabled in two ways:
- Manually by checking ‘Large context’ from Settings > Features
- Automatically when a Chat session grows long and/or includes large file attachments; in this case, large context charges will be indicated by a ⚠︎ icon appearing in the input interface
You can monitor per requests costs in real time from the Settings page.
Context window sizes
Mode | Context Window |
---|---|
Agent: claude-3.7-sonnet MAX | 200,000 tokens |
Agent: claude-3.7-sonnet | 120,000 tokens |
Agent: Other Models | 60,000 tokens |
⌘K | 10,000 tokens |
These thresholds are subject to change as Cursor further optimizes its context capabilities.
Model hosting
Models are hosted on US-based infrastructure by the model’s provider, a trusted partner or Cursor.
When Privacy Mode is enabled from Settings, Cursor nor the model providers will store your data with all data deleted after each request is processed. For further details see our Privacy, Privacy Policy, and Security pages.
Was this page helpful?