Dragonfly
Dragonfly serves as a seamless substitute for Redis, offering enhanced performance while reducing costs. It is specifically engineered to harness the capabilities of contemporary cloud infrastructure, catering to the data requirements of today’s applications, thereby liberating developers from the constraints posed by conventional in-memory data solutions. Legacy software cannot fully exploit the advantages of modern cloud technology. With its optimization for cloud environments, Dragonfly achieves an impressive 25 times more throughput and reduces snapshotting latency by 12 times compared to older in-memory data solutions like Redis, making it easier to provide the immediate responses that users demand. The traditional single-threaded architecture of Redis leads to high expenses when scaling workloads. In contrast, Dragonfly is significantly more efficient in both computation and memory usage, potentially reducing infrastructure expenses by up to 80%. Initially, Dragonfly scales vertically, only transitioning to clustering when absolutely necessary at a very high scale, which simplifies the operational framework and enhances system reliability. Consequently, developers can focus more on innovation rather than infrastructure management.
Learn more
RaimaDB
RaimaDB, an embedded time series database that can be used for Edge and IoT devices, can run in-memory. It is a lightweight, secure, and extremely powerful RDBMS. It has been field tested by more than 20 000 developers around the world and has been deployed in excess of 25 000 000 times.
RaimaDB is a high-performance, cross-platform embedded database optimized for mission-critical applications in industries such as IoT and edge computing. Its lightweight design makes it ideal for resource-constrained environments, supporting both in-memory and persistent storage options. RaimaDB offers flexible data modeling, including traditional relational models and direct relationships through network model sets. With ACID-compliant transactions and advanced indexing methods like B+Tree, Hash Table, R-Tree, and AVL-Tree, it ensures data reliability and efficiency. Built for real-time processing, it incorporates multi-version concurrency control (MVCC) and snapshot isolation, making it a robust solution for applications demanding speed and reliability.
Learn more
TensorWave
TensorWave is a cloud platform designed for AI and high-performance computing (HPC), exclusively utilizing AMD Instinct Series GPUs to ensure optimal performance. It features a high-bandwidth and memory-optimized infrastructure that seamlessly scales to accommodate even the most rigorous training or inference tasks. Users can access AMD’s leading GPUs in mere seconds, including advanced models like the MI300X and MI325X, renowned for their exceptional memory capacity and bandwidth, boasting up to 256GB of HBM3E and supporting speeds of 6.0TB/s. Additionally, TensorWave's architecture is equipped with UEC-ready functionalities that enhance the next generation of Ethernet for AI and HPC networking, as well as direct liquid cooling systems that significantly reduce total cost of ownership, achieving energy cost savings of up to 51% in data centers. The platform also incorporates high-speed network storage, which provides transformative performance, security, and scalability for AI workflows. Furthermore, it ensures seamless integration with a variety of tools and platforms, accommodating various models and libraries to enhance user experience. TensorWave stands out for its commitment to performance and efficiency in the evolving landscape of AI technology.
Learn more
TranslateGemma
TranslateGemma is an innovative collection of open machine translation models created by Google, based on the Gemma 3 architecture, which facilitates communication between individuals and systems in 55 languages by providing high-quality AI translations while ensuring efficiency and wide deployment options. Offered in sizes of 4 B, 12 B, and 27 B parameters, TranslateGemma encapsulates sophisticated multilingual functionalities into streamlined models that are capable of functioning on mobile devices, consumer laptops, local systems, or cloud infrastructure, all without compromising on precision or performance; assessments indicate that the 12 B variant can exceed the capabilities of larger baseline models while requiring less computational power. The development of these models involved a distinct two-phase fine-tuning approach that integrates high-quality human and synthetic translation data, using reinforcement learning to enhance translation accuracy across a variety of language families. This innovative methodology ensures that users benefit from an array of languages while experiencing swift and reliable translations.
Learn more