Skip to main content
Glama

kimi-latest-128k vs devstral-small-2507

Pricing, Performance & Features Comparison

Price unit:
Authormoonshot
Context Length128K
Reasoning
-
Providers1
ReleasedJul 2025
Knowledge Cutoff-
License-

Kimi-latest-128k refers to the Kimi K2 model, a state-of-the-art Mixture-of-Experts (MoE) language model with 32 billion activated and 1 trillion total parameters. It features a 128K context length and is meticulously optimized for agentic capabilities, specifically designed for tool use, reasoning, and autonomous problem-solving.

Input$2
Output$5
Latency (p50)2.4s
Output Limit128K
Function Calling
JSON Mode
-
InputText, Image, Audio, Video
OutputText, Audio
in$2out$5cache$0.15-
Latency (24h)
Success Rate (24h)
Authormistral
Context Length128K
Reasoning
-
Providers1
ReleasedJul 2025
Knowledge Cutoff-
License-

Devstral-Small-2507 is an agentic Large Language Model (LLM) developed by Mistral AI and All Hands AI specifically for software engineering tasks. It excels at using tools to explore codebases, edit multiple files, and power software engineering agents. This model is notable for its remarkable performance on the SWE-bench benchmark, where it is positioned as the #1 open-source model.

Input$0.1
Output$0.3
Latency (p50)639ms
Output Limit128K
Function Calling
JSON Mode
InputText
OutputText
in$0.1out$0.3--
Latency (24h)
Success Rate (24h)