deepseek-r1-distill-qwen-32b vs deepseek-r1-distill-llama-70b
Pricing, Performance & Features Comparison
DeepSeek-R1-Distill-Qwen-32B outperforms OpenAI-o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.
Input$0.45
Output$0.7
Latency (p50)8.2s
Output Limit8K
Function Calling
-
JSON Mode
-
InputText
OutputText
Latency (24h)
Success Rate (24h)
DeepSeek-R1-Distill-Llama-70B is a highly efficient language model that leverages knowledge distillation to achieve state-of-the-art performance. This model distills the reasoning patterns of larger models into a smaller, more agile architecture, resulting in exceptional results on benchmarks like AIME 2024, MATH-500, and LiveCodeBench. With 70 billion parameters, DeepSeek-R1-Distill-Llama-70B offers a unique balance of accuracy and efficiency, making it an ideal choice for a wide range of natural language processing tasks.
Input$0.55
Output$2.2
Latency (p50)-
Output Limit8K
Function Calling
-
JSON Mode
-
InputText
OutputText
in$0.55out$2.2--