Skip to main content
Glama

ministral-8b-2410 vs qwen-2.5-72b-instruct

Pricing, Performance & Features Comparison

Price unit:
Authormistral
Context Length128K
Reasoning
-
Providers1
ReleasedSep 2024
Knowledge CutoffOct 2023
License-

Ministral-8B-Instruct-2410 is an instruction-tuned language model built on Mistral’s 8B-parameter dense transformer architecture. It supports large context windows (up to 128k tokens) and is particularly strong in multilingual applications, code-related tasks, and chat-based interactions. Its design targets efficient on-device and edge computing scenarios with high performance at scale.

Input$0.1
Output$0.1
Latency (p50)674ms
Output Limit128K
Function Calling
JSON Mode
-
InputText
OutputText
in$0.1out$0.1--
Latency (24h)
Success Rate (24h)
Authoralibaba
Context Length131K
Reasoning
-
Providers1
ReleasedSep 2024
Knowledge Cutoff-
License-

Qwen2.5-72B-Instruct is a 72-billion-parameter, decoder-only language model designed for advanced instruction following and long-text generation. It excels at structured data understanding and output, especially JSON, and offers improved coding and mathematical reasoning. The model also supports over 29 languages and can handle extended contexts of up to 128K tokens.

Input$0.23
Output$0.4
Latency (p50)-
Output Limit8K
Function Calling
-
JSON Mode
InputText
OutputText
in$0.23out$0.4--