Max Mode gives you access to the full capabilities of Cursor’s advanced AI models. It’s designed for moments when you need additional processing power and deeper analysis.

Comparison

The main difference between normal mode and Max mode is context behaviour where Max mode is optimized to work through as much context as possible. In practice, this means:

  • Larger context windows
  • Up to 200 tool calls (without asking for continuation)
  • Read file tool can read up to 750 lines

Context Window Comparison

Here’s what different context window sizes can handle in practice, with examples from real-world codebases:

TokensScaleReal-World ExamplesWhat Fits
10,000SmallSingle utility librariesA utility like Underscore.js, or a few React components
60,000MediumUtility collectionsMost of a medium-sized library like Lodash
120,000LargeFull librariesComplete utility libraries or core parts of larger frameworks
200,000Extra LargeWeb frameworksFull frameworks like Express, or runtime libraries like Tokio
1,000,000MassiveFramework coresCore of major frameworks like Django (without tests)

Real-world token usage varies based on code style, comments, and tokenization. For example, Google’s documentation suggests that 1M tokens typically handles ~30,000-50,000 lines of code at 80 characters per line in practice.

Pricing

Max mode is offered at API cost, just as if you’d have brought your own key (BYOK). For detailed pricing, see models list

Token usage from your account dashboard.

Token types

Model providers offer pricing in different tiers, depending on capabilities for caching etc. Here’s a breakdown of the various token types and what they mean. You can see an example of how tokens are determined using OpenAI’s tokenizer (note that this is just for demonstration - we use different tokenization under the hood).

TypeDescriptionExampleCost
InputTokens the model is seeing for the first time,New messagesExpensive
Input (Cached)Tokens the model has seen before and have decided to cache for future userPrevious messages in a chatCheap
OutputThe returned tokens by the modelModel response, mostly codeExpensive

See full pricing in models

How to use Max mode

To use Max mode,

  1. Open model picker
  2. Toggle Max mode
  3. Select a compatible model from the dropdown

FAQ

When should I use Max mode?

Max Mode works particularly well for the hardest problems where you need the model to think and reason the most. Normal mode is still recommended for most tasks.

Why don’t all models have Max mode?

Max mode is intended for the large, context intensive operations. Models with context windows of 100k tokens or less see diminishing returns with Max mode