Skip to main content
Glama

deepseek-r1-distill-qwen-32b vs mistral-7b-instruct

Pricing, Performance & Features Comparison

Price unit:
Authordeepseek
Context Length128K
Reasoning
-
Providers3
ReleasedJan 2015
Knowledge CutoffJul 2024
License-

DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.

Input$0.45
Output$0.7
Latency (p50)6.9s
Output Limit8K
Function Calling
-
JSON Mode
-
InputText
OutputText
glama
Cheapest
in$0.45out$1.6--
in$0.5out$4.9--
in$0.7out$0.7--
Latency (24h)
Success Rate (24h)
Authormistral
Context Length32K
Reasoning
-
Providers1
ReleasedSep 2023
Knowledge Cutoff-
License-

The mistralai/mistral-7b-instruct series is a 7B-parameter language model fine-tuned for instruction-based tasks. It supports an extended context window (up to 32K tokens) and can handle function calling, demonstrating strong instruct performance. As an early demonstration, it lacks built-in content moderation mechanisms.

Input$0.03
Output$0.055
Latency (p50)-
Output Limit256
Function Calling
JSON Mode
-
InputText
OutputText
in$0.03out$0.055--