Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
In honor of Archimedes, whose 2311th anniversary we celebrate this year, we are excited to introduce our inaugural Mathstral model, a specialized 7B architecture tailored for mathematical reasoning and scientific exploration. This model features a 32k context window and is released under the Apache 2.0 license. Our intention behind contributing Mathstral to the scientific community is to enhance the pursuit of solving advanced mathematical challenges that necessitate intricate, multi-step logical reasoning. The launch of Mathstral is part of our wider initiative to support academic endeavors, developed in conjunction with Project Numina. Much like Isaac Newton during his era, Mathstral builds upon the foundation laid by Mistral 7B, focusing on STEM disciplines. It demonstrates top-tier reasoning capabilities within its category, achieving remarkable results on various industry-standard benchmarks. Notably, it scores 56.6% on the MATH benchmark and 63.47% on the MMLU benchmark, showcasing the performance differences by subject between Mathstral 7B and its predecessor, Mistral 7B, further emphasizing the advancements made in mathematical modeling. This initiative aims to foster innovation and collaboration within the mathematical community.
Description
Sarvam-105B stands as the premier large language model within Sarvam’s open-source lineup, engineered to provide exceptional reasoning capabilities, multilingual comprehension, and agent-driven execution all within a unified and scalable framework. This Mixture-of-Experts (MoE) model boasts an impressive total of approximately 105 billion parameters, activating only a subset for each token, which allows it to maintain superior computational efficiency while excelling in intricate tasks. It is particularly optimized for advanced reasoning, programming, mathematical challenges, and agentic processes, positioning it well for scenarios that necessitate multi-step problem-solving and organized outputs rather than merely engaging in basic conversations. With the ability to process long contexts of around 128K tokens, Sarvam-105B can effectively manage extensive documents, prolonged discussions, and complex analytical inquiries, ensuring coherence throughout. Additionally, its design facilitates a diverse range of applications, providing users with versatile tools to tackle a variety of intellectual challenges.
API Access
Has API
API Access
Has API
Integrations
Hugging Face
AI-FLOW
Airtrain
AlphaCorp
Fleak
GaiaNet
Graydient AI
HoneyHive
Literal AI
Lunary
Integrations
Hugging Face
AI-FLOW
Airtrain
AlphaCorp
Fleak
GaiaNet
Graydient AI
HoneyHive
Literal AI
Lunary
Pricing Details
Free
Free Trial
Free Version
Pricing Details
Free
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
Mistral AI
Founded
2023
Country
France
Website
mistral.ai/news/mathstral/
Vendor Details
Company Name
Sarvam
Founded
2023
Country
India
Website
www.sarvam.ai/blogs/sarvam-30b-105b