Enabling Auto configures Cursor to select the premium model best fit for the immediate task and with the highest reliability based on current demand. This feature can detect degraded output performance and automatically switch models to resolve it.
A context window is the maximum span of tokens (text and code) an LLM can consider at once, including both the input prompt and output generated by the model.
Each chat in Cursor maintains its own context window. The more prompts, attached files, and responses included in a session, the larger the context window grows.
Normally, Cursor uses a context window of 128k tokens (~10,000 lines of code). Max Mode is our option to turn on the maximum context windows for all models. This will be a bit slower and more expensive. It is most relevant for Gemini 2.5 Pro and GPT 4.1, which have 1M token context windows.
Models are hosted on US-based infrastructure by the model’s provider, a trusted partner or Cursor.
When Privacy Mode is enabled, neither Cursor nor model providers store your data. All data is deleted after each request. For details see our Privacy, Privacy Policy, and Security pages.
Assistant
Responses are generated using AI and may contain mistakes.
Enabling Auto configures Cursor to select the premium model best fit for the immediate task and with the highest reliability based on current demand. This feature can detect degraded output performance and automatically switch models to resolve it.
A context window is the maximum span of tokens (text and code) an LLM can consider at once, including both the input prompt and output generated by the model.
Each chat in Cursor maintains its own context window. The more prompts, attached files, and responses included in a session, the larger the context window grows.
Normally, Cursor uses a context window of 128k tokens (~10,000 lines of code). Max Mode is our option to turn on the maximum context windows for all models. This will be a bit slower and more expensive. It is most relevant for Gemini 2.5 Pro and GPT 4.1, which have 1M token context windows.
Models are hosted on US-based infrastructure by the model’s provider, a trusted partner or Cursor.
When Privacy Mode is enabled, neither Cursor nor model providers store your data. All data is deleted after each request. For details see our Privacy, Privacy Policy, and Security pages.
Assistant
Responses are generated using AI and may contain mistakes.