by Michael Truell
Instead of being a 'wrapper', Cursor combines large foundation models with small, custom-trained models. Small models handle high-frequency, low-latency tasks (like next-edit prediction), while large models handle complex reasoning, creating a seamless user experience.
Core Principles
- 1.Specialized Small Models: Train custom models for specific tasks (e.g., diff prediction) to achieve <300ms latency.
- 2.Input/Output Filtering: Use models to pre-search the codebase (context retrieval) and post-process output into valid code diffs.
- 3.Cost Arbitrage: Offload repetitive inference to cheaper, faster models to maintain economic viability.
- +1 more...
"At this point, every magic moment in Cursor involves a custom model in some way."