Mixtral 8x7B Description

Mixtral 8x7B has open weights and is a high quality sparse mixture expert model (SMoE). Licensed under Apache 2.0. Mixtral outperforms Llama 70B in most benchmarks, with 6x faster Inference. It is the strongest model with an open-weight license and the best overall model in terms of cost/performance tradeoffs. It matches or exceeds GPT-3.5 in most standard benchmarks.

Pricing

Pricing Starts At:
Free
Pricing Information:
Open source
Free Version:
Yes

Integrations

Reviews

Total
ease
features
design
support

No User Reviews. Be the first to provide a review:

Write a Review

Company Details

Company:
Mistral AI
Year Founded:
2023
Headquarters:
France
Website:
mistral.ai/news/mixtral-of-experts/

Media

Mixtral 8x7B Screenshot 1
Recommended Products
Top Rated Business VoIP Provider for 2024 for as low as $20/mo*! Icon
Top Rated Business VoIP Provider for 2024 for as low as $20/mo*!

Message, video, and phone on any device. Trusted by over 400,000 businesses.

- Includes 100+ Premium Features
- Unlimited Calling, Faxing, SMS, Conferencing.

Product Details

Platforms
Windows
Mac
Linux
On-Premises
Type of Training
Documentation

Mixtral 8x7B Features and Options