RaimaDB
RaimaDB, an embedded time series database that can be used for Edge and IoT devices, can run in-memory. It is a lightweight, secure, and extremely powerful RDBMS. It has been field tested by more than 20 000 developers around the world and has been deployed in excess of 25 000 000 times.
RaimaDB is a high-performance, cross-platform embedded database optimized for mission-critical applications in industries such as IoT and edge computing. Its lightweight design makes it ideal for resource-constrained environments, supporting both in-memory and persistent storage options. RaimaDB offers flexible data modeling, including traditional relational models and direct relationships through network model sets. With ACID-compliant transactions and advanced indexing methods like B+Tree, Hash Table, R-Tree, and AVL-Tree, it ensures data reliability and efficiency. Built for real-time processing, it incorporates multi-version concurrency control (MVCC) and snapshot isolation, making it a robust solution for applications demanding speed and reliability.
Learn more
Dragonfly
Dragonfly serves as a seamless substitute for Redis, offering enhanced performance while reducing costs. It is specifically engineered to harness the capabilities of contemporary cloud infrastructure, catering to the data requirements of today’s applications, thereby liberating developers from the constraints posed by conventional in-memory data solutions. Legacy software cannot fully exploit the advantages of modern cloud technology. With its optimization for cloud environments, Dragonfly achieves an impressive 25 times more throughput and reduces snapshotting latency by 12 times compared to older in-memory data solutions like Redis, making it easier to provide the immediate responses that users demand. The traditional single-threaded architecture of Redis leads to high expenses when scaling workloads. In contrast, Dragonfly is significantly more efficient in both computation and memory usage, potentially reducing infrastructure expenses by up to 80%. Initially, Dragonfly scales vertically, only transitioning to clustering when absolutely necessary at a very high scale, which simplifies the operational framework and enhances system reliability. Consequently, developers can focus more on innovation rather than infrastructure management.
Learn more
HEAVY.AI
HEAVY.AI is a pioneer in accelerated analysis. The HEAVY.AI platform can be used by government and business to uncover insights in data that is beyond the reach of traditional analytics tools. The platform harnesses the huge parallelism of modern CPU/GPU hardware and is available both in the cloud or on-premise. HEAVY.AI was developed from research at Harvard and MIT Computer Science and Artificial Intelligence Laboratory. You can go beyond traditional BI and GIS and extract high-quality information from large datasets with no lag by leveraging modern GPU and CPU hardware. To get a complete picture of what, when and where, unify and explore large geospatial or time-series data sets. Combining interactive visual analytics, hardware accelerated SQL, advanced analytics & data sciences frameworks, you can find the opportunity and risk in your enterprise when it matters most.
Learn more
Apache Doris
Apache Doris serves as a cutting-edge data warehouse tailored for real-time analytics, enabling exceptionally rapid analysis of data at scale.
It features both push-based micro-batch and pull-based streaming data ingestion that occurs within a second, alongside a storage engine capable of real-time upserts, appends, and pre-aggregation.
With its columnar storage architecture, MPP design, cost-based query optimization, and vectorized execution engine, it is optimized for handling high-concurrency and high-throughput queries efficiently.
Moreover, it allows for federated querying across various data lakes, including Hive, Iceberg, and Hudi, as well as relational databases such as MySQL and PostgreSQL.
Doris supports complex data types like Array, Map, and JSON, and includes a Variant data type that facilitates automatic inference for JSON structures, along with advanced text search capabilities through NGram bloomfilters and inverted indexes.
Its distributed architecture ensures linear scalability and incorporates workload isolation and tiered storage to enhance resource management.
Additionally, it accommodates both shared-nothing clusters and the separation of storage from compute resources, providing flexibility in deployment and management.
Learn more