Skip to main content
Glama

qwq-32b-preview vs mistral-large-2411

Pricing, Performance & Features Comparison

Price unit:
Authoralibaba
Context Length33K
Reasoning
-
Providers1
ReleasedNov 2024
Knowledge CutoffOct 2023
LicenseApache License 2.0

QwQ-32B-Preview is an experimental research model focusing on AI reasoning, with strong capabilities in math and coding. It features 32.5 billion parameters and a 32,768-token context window, leveraging transformer architecture with RoPE and advanced attention mechanisms. Despite its strengths, it has certain language mixing and reasoning limitations that remain areas of active research.

Input$0.17
Output$0.7
Latency (p50)13.6s
Output Limit512
Function Calling
-
JSON Mode
-
InputText, Image
OutputText
in$0.17out$0.7--
Latency (24h)
Success Rate (24h)
Authormistral
Context Length128K
Reasoning
-
Providers1
ReleasedNov 2024
Knowledge CutoffOct 2023
License-

Mistral Large 24.11 is a 123-billion-parameter language model designed for advanced reasoning, coding, and multilingual tasks. It supports a 128k context window with robust function-calling and JSON output capabilities. The model excels in complex reasoning scenarios, retrieval-augmented generation, and multi-format output generation.

Input$2
Output$6
Latency (p50)813ms
Output Limit4K
Function Calling
JSON Mode
InputText
OutputText
in$2out$6--
Latency (24h)
Success Rate (24h)