Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
Magistral is the inaugural language model family from Mistral AI that emphasizes reasoning, offered in two variants: Magistral Small, a 24 billion parameter open-weight model accessible under Apache 2.0 via Hugging Face, and Magistral Medium, a more robust enterprise-grade version that can be accessed through Mistral's API, the Le Chat platform, and various major cloud marketplaces. Designed for specific domains, it excels in transparent, multilingual reasoning across diverse tasks such as mathematics, physics, structured calculations, programmatic logic, decision trees, and rule-based systems, generating outputs that follow a chain of thought in the user's preferred language, which can be easily tracked and validated. This release signifies a transition towards more compact yet highly effective transparent AI reasoning capabilities. Currently, Magistral Medium is in preview on platforms including Le Chat, the API, SageMaker, WatsonX, Azure AI, and Google Cloud Marketplace. Its design is particularly suited for general-purpose applications that necessitate extended thought processes and improved accuracy compared to traditional non-reasoning language models. The introduction of Magistral represents a significant advancement in the pursuit of sophisticated reasoning in AI applications.
Description
Mistral AI has launched two cutting-edge models designed for on-device computing and edge applications, referred to as "les Ministraux": Ministral 3B and Ministral 8B. These innovative models redefine the standards of knowledge, commonsense reasoning, function-calling, and efficiency within the sub-10B category. They are versatile enough to be utilized or customized for a wide range of applications, including managing complex workflows and developing specialized task-focused workers. Capable of handling up to 128k context length (with the current version supporting 32k on vLLM), Ministral 8B also incorporates a unique interleaved sliding-window attention mechanism to enhance both speed and memory efficiency during inference. Designed for low-latency and compute-efficient solutions, these models excel in scenarios such as offline translation, smart assistants that don't rely on internet connectivity, local data analysis, and autonomous robotics. Moreover, when paired with larger language models like Mistral Large, les Ministraux can effectively function as streamlined intermediaries, facilitating function-calling within intricate multi-step workflows, thereby expanding their applicability across various domains. This combination not only enhances performance but also broadens the scope of what can be achieved with AI in edge computing.
API Access
Has API
API Access
Has API
Integrations
Le Chat
APIPark
DataChain
GMTech
Graydient AI
HumanLayer
Kiin
Langflow
Lunary
Mathstral
Integrations
Le Chat
APIPark
DataChain
GMTech
Graydient AI
HumanLayer
Kiin
Langflow
Lunary
Mathstral
Pricing Details
No price information available.
Free Trial
Free Version
Pricing Details
Free
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
Mistral AI
Founded
2023
Country
France
Website
mistral.ai/news/magistral
Vendor Details
Company Name
Mistral AI
Founded
2023
Country
France
Website
mistral.ai/news/ministraux/