Cursor supports all frontier coding models from all major model providers.

Context windows

A context window is the maximum span of tokens (text and code) an LLM can consider at once, including both the input prompt and output generated by the model.

Each chat in Cursor maintains its own context window. The more prompts, attached files, and responses included in a session, the larger the context window grows.

Learn more about working with context in Cursor.

Max Mode

Normally, Cursor uses a context window of 128k tokens (~10,000 lines of code). Max Mode is our option to turn on the maximum context windows for all models. This will be a bit slower and more expensive. It is most relevant for Gemini 2.5 Pro and GPT 4.1, which have 1M token context windows.

Auto

Enabling Auto configures Cursor to select the premium model best fit for the immediate task and with the highest reliability based on current demand. This feature can detect degraded output performance and automatically switch models to resolve it.

FAQ