Kimi K2 is our latest Mixture-of-Experts model with 32 billion activated parameters and 1 trillion total parameters. It achieves state-of-the-art performance in frontier knowledge, math, and coding among non-thinking models. But it goes further — meticulously optimized for agentic tasks, Kimi K2 does not just answer; it acts. And now, it is within your reach. Today, we are open-sourcing: Kimi-K2-Base: The foundation model, a strong start for researchers and builders who want full control for fine-tuning and custom solutions. Kimi-K2-Instruct: The post-trained model best for drop-in, general-purpose chat and agentic experiences. It is a reflex-grade model without long thinking. With Kimi K2, advanced agentic intelligence is more open and accessible than ever. We can't wait to see what you build.
Notes: 6 FLOP / parameter / token * 32 * 10^9 activated parameters * 15.5 * 10^12 tokens = 2.976e+24 FLOP
Size Notes: "Kimi K2 was pre-trained on 15.5T tokens"
Notes: MoE with 1T total parameters and 32B parameters active per forward pass