Mixtral 8x7B
ActiveOpen-weights mixture-of-experts model with GPT-3.5 class performance.
Overview
Mixtral 8x7B uses a sparse mixture-of-experts architecture (8 experts, 2 active per token) to deliver GPT-3.5 class performance at lower inference cost than a dense 45B model. Apache-2.0 licensed.
Integrations & tooling support
- Tool calling
- Supported
- Structured outputs
- Not supported
Price vs quality
Not enough data
This model has no benchmark scores recorded yet.
Community ratings
No ratings yet. Be the first to rate Mixtral 8x7B.
Rate Mixtral 8x7B
Sign in to rate and review.
Comments
Sign in to leave a comment.