Skip to main content
Glama

moonshot-v1-128k vs open-mixtral-8x7b

Pricing, Performance & Features Comparison

Price unit:
Authormoonshot
Context Length128K
Reasoning
-
Providers1
ReleasedJan 2023
Knowledge Cutoff-
License-

Moonshot-v1-128k is a large language model with ultra-long context processing capabilities, capable of handling up to 128,000 tokens. It is designed for generating extremely long texts and meeting the demands of complex generation tasks, making it ideal for research, academia, and large document generation.

Input$2
Output$5
Latency (p50)1.2s
Output Limit128K
Function Calling
JSON Mode
-
InputText
OutputText
in$2out$5--
Latency (24h)
Success Rate (24h)
Authormistral
Context Length33K
Reasoning
-
Providers1
ReleasedFeb 2024
Knowledge CutoffDec 2023
License-

Mixtral 8x7B is a high-quality sparse mixture of experts (SMoE) large language model with open weights, released under Apache 2.0 license. Despite having 45 billion parameters, its architecture requires compute equivalent to a 14 billion parameter model, enabling 6x faster inference and strong performance that outperforms Llama 2 70B and matches or exceeds GPT-3.5 on many benchmarks.

Input$0.7
Output$0.7
Latency (p50)744ms
Output Limit4K
Function Calling
JSON Mode
InputText
OutputText
in$0.7out$0.7--
Latency (24h)
Success Rate (24h)