Molmo represents a cutting-edge family of multimodal AI models crafted by the Allen Institute for AI (Ai2). These innovative models are specifically engineered to connect the divide between open-source and proprietary systems, ensuring they perform competitively across numerous academic benchmarks and assessments by humans. In contrast to many existing multimodal systems that depend on synthetic data sourced from proprietary frameworks, Molmo is exclusively trained on openly available data, which promotes transparency and reproducibility in AI research. A significant breakthrough in the development of Molmo is the incorporation of PixMo, a unique dataset filled with intricately detailed image captions gathered from human annotators who utilized speech-based descriptions, along with 2D pointing data that empowers the models to respond to inquiries with both natural language and non-verbal signals. This capability allows Molmo to engage with its surroundings in a more sophisticated manner, such as by pointing to specific objects within images, thereby broadening its potential applications in diverse fields, including robotics, augmented reality, and interactive user interfaces. Furthermore, the advancements made by Molmo set a new standard for future multimodal AI research and application development.