Skip to main content
Cursor's Composer 2, built on a Chinese AI model, highlights Western open-source vulnerabilities.

Editorial illustration for Cursor's Composer 2 built on Chinese AI model, exposing Western open-source gaps

Cursor's AI Composer Reveals Open-Source Model Secrets

Cursor's Composer 2 built on Chinese AI model, exposing Western open-source gaps

2 min read

Cursor’s latest release, Composer 2, quietly runs on a Chinese‑origin model, a fact that slipped past most reviewers until a deep dive of the code revealed the truth. The revelation has sparked a broader conversation about why Western open‑source projects still lean on heavyweight proprietary engines for everyday reasoning tasks. Developers praised the open‑source community’s push toward self‑sufficiency, yet the gap between ambition and execution remains stark.

In the wake of the discovery, analysts are asking whether the community’s flagship models can truly match the performance of commercial alternatives without resorting to complex architectures that hide their inner workings. The answer, they argue, lies in the details of how these models are built and how many parameters they actually engage when generating a response. That brings us to a key observation about gpt‑oss‑120b, a model many have hailed as a milestone for open‑source AI.

While gpt-oss-120b is a monumental achievement for Western open source--offering reasoning capabilities that rival proprietary models like o4-mini--it is fundamentally a sparse Mixture-of-Experts (MoE) model that activates only 5.1 billion parameters per token. For a general-purpose reasoning assistant, that is an efficiency masterstroke; for a tool like Composer 2, which must maintain structural coherence across a 256,000-token context window, it is arguably too "thin." By contrast, Kimi K2.5 is a 1-trillion-parameter titan that keeps 32 billion parameters active at any given moment. In the high-stakes world of agentic coding, sheer cognitive mass still dictates performance, and Cursor clearly calculated that Kimi's 6x advantage in active parameter count was essential for synthesizing the "context explosion" that occurs during complex, multi-step autonomous programming tasks.

The reveal was unexpected. Cursor's $29.3 billion Composer 2, marketed as a frontier‑level coding assistant, actually runs on Moonshot AI's Kimi K2.5, a Chinese open‑source model. By omitting that provenance, the company positioned itself as a research‑driven lab rather than an IDE wrapper, a distinction that now appears blurred.

The episode highlights a gap: Western open‑source projects such as gpt‑oss‑120b, while praised for reasoning that rivals proprietary offerings, rely on sparse Mixture‑of‑Experts architectures that activate only 5.1 billion parameters per token. Whether this technical difference translates into a lasting advantage for Western tools remains uncertain. Critics note the reliance on a foreign model may expose users to supply‑chain opacity, yet supporters say it's still meeting developer needs.

As the market watches, the broader question of how open‑source ecosystems can match the resources behind large commercial ventures stays open. Ultimately, the incident underscores the importance of transparency in model provenance for any AI‑driven development product.

Further Reading

Common Questions Answered

What Chinese AI model does Cursor's Composer 2 actually run on?

Composer 2 runs on Moonshot AI's Kimi K2.5, a Chinese open-source model that was not initially disclosed by Cursor. This revelation has raised questions about the transparency of the company's AI technology and marketing approach.

Why are Western open-source AI projects struggling to compete with proprietary models?

Western open-source projects like gpt-oss-120b use sparse Mixture-of-Experts (MoE) models that activate only a fraction of their parameters per token, which can limit their effectiveness for complex tasks. This approach, while efficient, may not provide the depth of reasoning and context maintenance required for advanced applications like Composer 2.

How does the Kimi K2.5 model compare to Western open-source AI models?

The Kimi K2.5 model appears to offer more robust performance for tasks requiring extensive context maintenance, such as maintaining structural coherence across a 256,000-token window. In contrast, Western models like gpt-oss-120b activate only 5.1 billion parameters per token, which may be less suitable for complex reasoning tasks.