Average Ratings 0 Ratings
Average Ratings 0 Ratings
Description
Molmo 2 represents a cutting-edge suite of open vision-language models that come with completely accessible weights, training data, and code, thereby advancing the original Molmo series' capabilities in grounded image comprehension to encompass video and multiple image inputs. This evolution enables sophisticated video analysis, including pointing, tracking, dense captioning, and question-answering functionalities, all of which demonstrate robust spatial and temporal reasoning across frames. The suite consists of three distinct models: an 8 billion-parameter variant tailored for comprehensive video grounding and QA tasks, a 4 billion-parameter model that prioritizes efficiency, and a 7 billion-parameter model backed by Olmo, which features a fully open end-to-end architecture that includes the foundational language model. Notably, these new models surpass their predecessors on key benchmarks, setting unprecedented standards for open-model performance in image and video comprehension tasks. Furthermore, they often rival significantly larger proprietary systems while being trained on a much smaller dataset compared to similar closed models, showcasing their efficiency and effectiveness in the field. This impressive achievement marks a significant advancement in the accessibility and performance of AI-driven visual understanding technologies.
Description
Olmo 3 represents a comprehensive family of open models featuring variations with 7 billion and 32 billion parameters, offering exceptional capabilities in base performance, reasoning, instruction, and reinforcement learning, while also providing transparency throughout the model development process, which includes access to raw training datasets, intermediate checkpoints, training scripts, extended context support (with a window of 65,536 tokens), and provenance tools. The foundation of these models is built upon the Dolma 3 dataset, which comprises approximately 9 trillion tokens and utilizes a careful blend of web content, scientific papers, programming code, and lengthy documents; this thorough pre-training, mid-training, and long-context approach culminates in base models that undergo post-training enhancements through supervised fine-tuning, preference optimization, and reinforcement learning with accountable rewards, resulting in the creation of the Think and Instruct variants. Notably, the 32 billion Think model has been recognized as the most powerful fully open reasoning model to date, demonstrating performance that closely rivals that of proprietary counterparts in areas such as mathematics, programming, and intricate reasoning tasks, thereby marking a significant advancement in open model development. This innovation underscores the potential for open-source models to compete with traditional, closed systems in various complex applications.
API Access
Has API
API Access
Has API
Integrations
Ai2 OLMoE
Bluesky
Hugging Face
Olmo 2
Threads
Pricing Details
No price information available.
Free Trial
Free Version
Pricing Details
Free
Free Trial
Free Version
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Deployment
Web-Based
On-Premises
iPhone App
iPad App
Android App
Windows
Mac
Linux
Chromebook
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Customer Support
Business Hours
Live Rep (24/7)
Online Support
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Types of Training
Training Docs
Webinars
Live Training (Online)
In Person
Vendor Details
Company Name
Ai2
Founded
2014
Country
United States
Website
allenai.org/blog/molmo2
Vendor Details
Company Name
Ai2
Founded
2014
Country
United States
Website
allenai.org/blog/olmo3