MoonshotAI: Kimi K2 Thinking
Kimi K2 Thinking is Moonshot AI’s most advanced open reasoning model to date, extending the K2 series into agentic, long-horizon reasoning. Built on the trillion-parameter Mixture-of-Experts (MoE) architecture introduced in Kimi K2, it activates 32 billion parameters per forward pass and supports 256 k-token context windows. The model is optimized for persistent step-by-step thought, dynamic tool invocation, and complex reasoning workflows that span hundreds of turns. It interleaves step-by-step reasoning with tool use, enabling autonomous research, coding, and writing that can persist for hundreds of sequential actions without drift. It sets new open-source benchmarks on HLE, BrowseComp, SWE-Multilingual, and LiveCodeBench, while maintaining stable multi-agent behavior through 200–300 tool calls. Built on a large-scale MoE architecture with MuonClip optimization, it combines strong reasoning depth with high inference efficiency for demanding agentic and analytical tasks.
Pricing per 1M Tokens
| Input (Prompt) | $0.47 |
| Output (Completion) | $2.00 |
| Cache Read | $0.14 |
| Cache Write | Free |
| Image | N/A |
Specifications
| Context Length | 131K |
| Max Output Tokens | N/A |
| Input Modalities | Text |
| Output Modalities | Text |
| Tokenizer | Other |
| Instruct Type | N/A |
| Top Provider Context | 131K |
| Top Provider Max Output | N/A |
| Moderated | No |
Compare this model
See how MoonshotAI: Kimi K2 Thinking stacks up against other models.
More from Moonshotai
Last updated: March 23, 2026
First tracked: March 23, 2026