glm-5 vs ministral-14b-2512
Pricing, Performance & Features Comparison
GLM-5 is a mixture-of-experts language model from Z.ai with 744 billion total parameters and 40 billion active parameters, designed for complex systems engineering and long-horizon agentic tasks. It utilizes DeepSeek Sparse Attention (DSA) to reduce deployment costs while maintaining long-context capacity, and achieves best-in-class performance among open-source models in reasoning, coding, and agentic tasks.
Input$1
Output$3.2
Latency (p50)6.8s
Output Limit131K
Function Calling
JSON Mode
InputText
OutputText
in$1out$3.2cache$0.2-
Latency (24h)
Success Rate (24h)
Ministral 3 14B is the largest model in the Ministral 3 family, offering state-of-the-art capabilities and performance comparable to its larger Mistral Small 3.2 24B counterpart. Optimized for local deployment, it delivers high performance across diverse hardware, including local setups.
Input$0.2
Output$0.2
Latency (p50)1.4s
Output Limit-
Function Calling
JSON Mode
-
Input-
Output-
in$0.2out$0.2--