Moonshot AI Unveils 1-Trillion Parameter Kimi-K2 Model
- •Moonshot AI releases Kimi-K2-Instruct-0905 featuring 1 trillion parameters using a Mixture-of-Experts architecture.
- •Model doubles its context window to 256,000 tokens to handle massive codebases and long-horizon tasks.
- •Kimi-K2 achieves 69.2% on SWE-Bench verified, rivaling top-tier proprietary models in autonomous coding.
Moonshot AI has officially unveiled Kimi-K2-Instruct-0905, a massive Mixture-of-Experts (MoE) model that signals a significant leap in large-scale language modeling. While the model contains a staggering 1 trillion total parameters, its architecture is designed for efficiency, activating only 32 billion parameters per token. This sparse activation allows for high-performance reasoning without the prohibitive computational costs typically associated with dense models of this scale.
The update focuses heavily on software engineering intelligence. By doubling its context window to 256,000 tokens, Kimi-K2 can ingest and reason over entire codebases, facilitating more coherent long-horizon programming and frontend design tasks. Benchmarks reflect this progress; the model posted a 69.2% success rate on SWE-Bench verified, positioning it as a formidable competitor to established industry leaders. It also introduces specialized optimizations for the aesthetics and functionality of generated code.
Beyond raw performance, Moonshot AI has prioritized ecosystem compatibility. Kimi-K2 is available via OpenAI and Anthropic-compatible APIs, streamlining the transition for developers already using existing workflows. With native support for 8-bit floating point (FP8) precision and integration with major inference engines like vLLM and Groq, the model is built for immediate, high-throughput deployment in production environments.