Best Oracle Cloud Infrastructure Streaming Alternatives in 2024
Find the top alternatives to Oracle Cloud Infrastructure Streaming currently available. Compare ratings, reviews, pricing, and features of Oracle Cloud Infrastructure Streaming alternatives in 2024. Slashdot lists the best Oracle Cloud Infrastructure Streaming alternatives on the market that offer competing products that are similar to Oracle Cloud Infrastructure Streaming. Sort through Oracle Cloud Infrastructure Streaming alternatives below to make the best choice for your needs
-
1
StarTree
StarTree
25 RatingsStarTree Cloud is a fully-managed real-time analytics platform designed for OLAP at massive speed and scale for user-facing applications. Powered by Apache Pinot, StarTree Cloud provides enterprise-grade reliability and advanced capabilities such as tiered storage, scalable upserts, plus additional indexes and connectors. It integrates seamlessly with transactional databases and event streaming platforms, ingesting data at millions of events per second and indexing it for lightning-fast query responses. StarTree Cloud is available on your favorite public cloud or for private SaaS deployment. StarTree Cloud includes StarTree Data Manager, which allows you to ingest data from both real-time sources such as Amazon Kinesis, Apache Kafka, Apache Pulsar, or Redpanda, as well as batch data sources such as data warehouses like Snowflake, Delta Lake or Google BigQuery, or object stores like Amazon S3, Apache Flink, Apache Hadoop, or Apache Spark. StarTree ThirdEye is an add-on anomaly detection system running on top of StarTree Cloud that observes your business-critical metrics, alerting you and allowing you to perform root-cause analysis — all in real-time. -
2
WarpStream
WarpStream
$2,987 per monthWarpStream, an Apache Kafka compatible data streaming platform, is built directly on object storage. It has no inter-AZ network costs, no disks that need to be managed, and it's infinitely scalable within your VPC. WarpStream is deployed in your VPC as a stateless, auto-scaling binary agent. No local disks are required to be managed. Agents stream data directly into and out of object storage without buffering on local drives and no data tiering. Instantly create new "virtual" clusters in our control plan. Support multiple environments, teams or projects without having to manage any dedicated infrastructure. WarpStream is Apache Kafka protocol compatible, so you can continue to use your favorite tools and applications. No need to rewrite or use a proprietary SDK. Simply change the URL of your favorite Kafka library in order to start streaming. Never again will you have to choose between budget and reliability. -
3
Striim
Striim
Data integration for hybrid clouds Modern, reliable data integration across both your private cloud and public cloud. All this in real-time, with change data capture and streams. Striim was developed by the executive and technical team at GoldenGate Software. They have decades of experience in mission critical enterprise workloads. Striim can be deployed in your environment as a distributed platform or in the cloud. Your team can easily adjust the scaleability of Striim. Striim is fully secured with HIPAA compliance and GDPR compliance. Built from the ground up to support modern enterprise workloads, whether they are hosted in the cloud or on-premise. Drag and drop to create data flows among your sources and targets. Real-time SQL queries allow you to process, enrich, and analyze streaming data. -
4
Informatica Data Engineering Streaming
Informatica
AI-powered Informatica Data Engineering streaming allows data engineers to ingest and process real-time streaming data in order to gain actionable insights. -
5
DeltaStream
DeltaStream
DeltaStream is an integrated serverless streaming processing platform that integrates seamlessly with streaming storage services. Imagine it as a compute layer on top your streaming storage. It offers streaming databases and streaming analytics along with other features to provide an integrated platform for managing, processing, securing and sharing streaming data. DeltaStream has a SQL-based interface that allows you to easily create stream processing apps such as streaming pipelines. It uses Apache Flink, a pluggable stream processing engine. DeltaStream is much more than a query-processing layer on top Kafka or Kinesis. It brings relational databases concepts to the world of data streaming, including namespacing, role-based access control, and enables you to securely access and process your streaming data, regardless of where it is stored. -
6
Google Cloud Dataflow
Google
Unified stream and batch data processing that is serverless, fast, cost-effective, and low-cost. Fully managed data processing service. Automated provisioning of and management of processing resource. Horizontal autoscaling worker resources to maximize resource use Apache Beam SDK is an open-source platform for community-driven innovation. Reliable, consistent processing that works exactly once. Streaming data analytics at lightning speed Dataflow allows for faster, simpler streaming data pipeline development and lower data latency. Dataflow's serverless approach eliminates the operational overhead associated with data engineering workloads. Dataflow allows teams to concentrate on programming and not managing server clusters. Dataflow's serverless approach eliminates operational overhead from data engineering workloads, allowing teams to concentrate on programming and not managing server clusters. Dataflow automates provisioning, management, and utilization of processing resources to minimize latency. -
7
Confluent
Confluent
Apache Kafka®, with Confluent, has an infinite retention. Be infrastructure-enabled, not infrastructure-restricted Legacy technologies require you to choose between being real-time or highly-scalable. Event streaming allows you to innovate and win by being both highly-scalable and real-time. Ever wonder how your rideshare app analyses massive amounts of data from multiple sources in order to calculate real-time ETA. Wondering how your credit card company analyzes credit card transactions from all over the world and sends fraud notifications in real time? Event streaming is the answer. Microservices are the future. A persistent bridge to the cloud can enable your hybrid strategy. Break down silos to demonstrate compliance. Gain real-time, persistent event transport. There are many other options. -
8
Axual
Axual
Axual provides Kafka-as-a-Service to DevOps teams. Our intuitive Kafka platform will empower your team to unlock insights, drive decisions and improve productivity. Axual is the ideal solution for enterprises that want to seamlessly integrate data streaming with their core IT infrastructure. Our all-in one Kafka platform was designed to eliminate the requirement for extensive technical skills or knowledge, and provide a ready-made product that delivers all of the benefits of event-streaming without the hassle. The Axual Platform, an all-in-one platform, is designed to simplify and enhance the deployment and management of Apache Kafka real-time streaming data. The Axual Platform offers a wide range of features to meet the needs of modern enterprises. This allows organizations to maximize the potential of data streaming, while minimizing complexity. -
9
Amazon MSK
Amazon
$0.0543 per hourAmazon MSK is a fully managed service that makes coding and running applications that use Apache Kafka for streaming data processing easy. Apache Kafka is an open source platform that allows you to build real-time streaming data applications and pipelines. Amazon MSK allows you to use native Apache Kafka APIs for populating data lakes, stream changes between databases, and to power machine learning or analytics applications. It is difficult to set up, scale, and manage Apache Kafka clusters in production. Apache Kafka clusters can be difficult to set up and scale on your own. -
10
Azure Event Hubs
Microsoft
$0.03 per hourEvent Hubs is a fully managed, real time data ingestion service that is simple, reliable, and scalable. Stream millions of events per minute from any source to create dynamic data pipelines that can be used to respond to business problems. Use the geo-disaster recovery or geo-replication features to continue processing data in emergencies. Integrate seamlessly with Azure services to unlock valuable insights. You can allow existing Apache Kafka clients to talk to Event Hubs with no code changes. This allows you to have a managed Kafka experience, without the need to manage your own clusters. You can experience real-time data input and microbatching in the same stream. Instead of worrying about infrastructure management, focus on gaining insights from your data. Real-time big data pipelines are built to address business challenges immediately. -
11
Apache Kafka
The Apache Software Foundation
1 RatingApache Kafka®, is an open-source distributed streaming platform. -
12
Rockset
Rockset
FreeReal-time analytics on raw data. Live ingest from S3, DynamoDB, DynamoDB and more. Raw data can be accessed as SQL tables. In minutes, you can create amazing data-driven apps and live dashboards. Rockset is a serverless analytics and search engine that powers real-time applications and live dashboards. You can directly work with raw data such as JSON, XML and CSV. Rockset can import data from real-time streams and data lakes, data warehouses, and databases. You can import real-time data without the need to build pipelines. Rockset syncs all new data as it arrives in your data sources, without the need to create a fixed schema. You can use familiar SQL, including filters, joins, and aggregations. Rockset automatically indexes every field in your data, making it lightning fast. Fast queries are used to power your apps, microservices and live dashboards. Scale without worrying too much about servers, shards or pagers. -
13
Amazon Kinesis
Amazon
You can quickly collect, process, analyze, and analyze video and data streams. Amazon Kinesis makes it easy for you to quickly and easily collect, process, analyze, and interpret streaming data. Amazon Kinesis provides key capabilities to process streaming data at any scale cost-effectively, as well as the flexibility to select the tools that best fit your application's requirements. Amazon Kinesis allows you to ingest real-time data, including video, audio, website clickstreams, application logs, and IoT data for machine learning, analytics, or other purposes. Amazon Kinesis allows you to instantly process and analyze data, rather than waiting for all the data to be collected before processing can begin. Amazon Kinesis allows you to ingest buffer and process streaming data instantly, so you can get insights in seconds or minutes, instead of waiting for hours or days. -
14
Apache Flink
Apache Software Foundation
Apache Flink is a distributed processing engine and framework for stateful computations using unbounded and bounded data streams. Flink can be used in all cluster environments and perform computations at any scale and in-memory speed. A stream of events can be used to produce any type of data. All data, including credit card transactions, machine logs, sensor measurements, and user interactions on a website, mobile app, are generated as streams. Apache Flink excels in processing both unbounded and bound data sets. Flink's runtime can run any type of application on unbounded stream streams thanks to its precise control of state and time. Bounded streams are internal processed by algorithms and data structure that are specifically designed to process fixed-sized data sets. This results in excellent performance. Flink can be used with all of the resource managers previously mentioned. -
15
IBM® Event Streams, an event-streaming platform built on Apache Kafka open-source software, is a smart app that reacts to events as they occur. Event Streams is based upon years of IBM operational experience running Apache Kafka stream events for enterprises. Event Streams is ideal for mission-critical workloads. You can extend the reach and reach of your enterprise assets by connecting to a variety of core systems and using a scalable RESTAPI. Disaster recovery is made easier by geo-replication and rich security. Use the CLI to take advantage of IBM productivity tools. Replicate data between Event Streams deployments during a disaster-recovery scenario.
-
16
Arroyo
Arroyo
Scale from 0 to millions of events every second. Arroyo is shipped as a single compact binary. Run locally on MacOS, Linux or Kubernetes for development and deploy to production using Docker or Kubernetes. Arroyo is an entirely new stream processing engine that was built from the ground-up to make real time easier than batch. Arroyo has been designed so that anyone with SQL knowledge can build reliable, efficient and correct streaming pipelines. Data scientists and engineers are able to build real-time dashboards, models, and applications from end-to-end without the need for a separate streaming expert team. SQL allows you to transform, filter, aggregate and join data streams with results that are sub-second. Your streaming pipelines should not page someone because Kubernetes rescheduled your pods. Arroyo can run in a modern, elastic cloud environment, from simple container runtimes such as Fargate, to large, distributed deployments using the Kubernetes logo. -
17
Cloudera DataFlow
Cloudera
You can manage your data from the edge to the cloud with a simple, no-code approach to creating sophisticated streaming applications. -
18
Redpanda
Redpanda Data
You can deliver customer experiences like never before with breakthrough data streaming capabilities Both the ecosystem and Kafka API are compatible. Redpanda BulletPredictable low latency with zero data loss. Redpanda BulletUp to 10x faster than Kafka Redpanda BulletEnterprise-grade support and hotfixes. Redpanda BulletAutomated backups for S3/GCS. Redpanda Bullet100% freedom of routine Kafka operations. Redpanda BulletSupports for AWS/GCP. Redpanda was built from the ground up to be easy to install and get running quickly. Redpanda's power will be evident once you have tried it in production. You can use the more advanced Redpanda functions. We manage all aspects of provisioning, monitoring, as well as upgrades. We do not have access to your cloud credentials. Sensitive data never leaves your environment. You can have it provisioned, operated, maintained, and updated for you. Configurable instance types. As your needs change, you can expand the cluster. -
19
Astra Streaming
DataStax
Responsive apps keep developers motivated and users engaged. With the DataStax Astra streaming service platform, you can meet these ever-increasing demands. DataStax Astra Streaming, powered by Apache Pulsar, is a cloud-native messaging platform and event streaming platform. Astra Streaming lets you build streaming applications on top a multi-cloud, elastically scalable and event streaming platform. Apache Pulsar is the next-generation event streaming platform that powers Astra Streaming. It provides a unified solution to streaming, queuing and stream processing. Astra Streaming complements Astra DB. Astra Streaming allows existing Astra DB users to easily create real-time data pipelines from and to their Astra DB instances. Astra Streaming allows you to avoid vendor lock-in by deploying on any major public cloud (AWS, GCP or Azure) compatible with open source Apache Pulsar. -
20
Spark Streaming
Apache Software Foundation
Spark Streaming uses Apache Spark's language-integrated API for stream processing. It allows you to write streaming jobs in the same way as you write batch jobs. It supports Java, Scala, and Python. Spark Streaming recovers lost work as well as operator state (e.g. Without any additional code, Spark Streaming recovers both lost work and operator state (e.g. sliding windows) right out of the box. Spark Streaming allows you to reuse the same code for batch processing and join streams against historical data. You can also run ad-hoc queries about stream state by running on Spark. Spark Streaming allows you to create interactive applications that go beyond analytics. Apache Spark includes Spark Streaming. It is updated with every Spark release. Spark Streaming can be run on Spark's standalone mode or other supported cluster resource mangers. It also has a local run mode that can be used for development. Spark Streaming uses ZooKeeper for high availability in production. -
21
Materialize
Materialize
$0.98 per hourMaterialize is a reactive database that provides incremental view updates. Our standard SQL allows developers to easily work with streaming data. Materialize connects to many external data sources without any pre-processing. Connect directly to streaming sources such as Kafka, Postgres databases and CDC or historical data sources such as files or S3. Materialize allows you to query, join, and transform data sources in standard SQL - and presents the results as incrementally-updated Materialized views. Queries are kept current and updated as new data streams are added. With incrementally-updated views, developers can easily build data visualizations or real-time applications. It is as easy as writing a few lines SQL to build with streaming data. -
22
IBM StreamSets
IBM
$1000 per monthIBM® StreamSets allows users to create and maintain smart streaming data pipelines using an intuitive graphical user interface. This facilitates seamless data integration in hybrid and multicloud environments. IBM StreamSets is used by leading global companies to support millions data pipelines, for modern analytics and intelligent applications. Reduce data staleness, and enable real-time information at scale. Handle millions of records across thousands of pipelines in seconds. Drag-and-drop processors that automatically detect and adapt to data drift will protect your data pipelines against unexpected changes and shifts. Create streaming pipelines for ingesting structured, semistructured, or unstructured data to deliver it to multiple destinations. -
23
Lenses
Lenses.io
$49 per monthAllow everyone to view and discover streaming data. Up to 95% of productivity can be increased by sharing, documenting, and cataloging data. Next, create apps for production use cases using the data. To address privacy concerns and cover all the gaps in open source technology, apply a data-centric security approach. Secure and low-code data pipeline capabilities. All darkness is eliminated and data and apps can be viewed with unparalleled visibility. Unify your data technologies and data meshes and feel confident using open source production. Independent third-party reviews have rated Lenses the best product for real time stream analytics. We have built features to allow you to focus on what is driving value from real-time data. This was based on feedback from our community as well as thousands of engineering hours. You can deploy and run SQL-based real-time applications over any Kafka Connect, Kubernetes or Kubernetes infrastructure, including AWS EKS. -
24
Leo
Leo
$251 per monthTransform your data into a live stream that is immediately available and ready for use. Leo makes event sourcing simpler by making it easy for you to create, visualize and monitor your data flows. You no longer have to be restricted by legacy systems once you unlock your data. Your developers and stakeholders will be happy with the dramatically reduced development time. Microservice architectures can be used to innovate and increase agility. Microservices are all about data. To make microservices a reality, an organization must have a reliable and repeatable backbone of data. Your custom app should support full-fledged searching. It won't be difficult to add and maintain a search database if you have the data. -
25
Solace PubSub+
Solace
Solace is a specialist in Event-Driven-Architecture (EDA), with two decades of experience providing enterprises with highly reliable, robust and scalable data movement technology based on the publish & subscribe (pub/sub) pattern. Solace technology enables the real-time data flow behind many of the conveniences you take for granted every day such as immediate loyalty rewards from your credit card, the weather data delivered to your mobile phone, real-time airplane movements on the ground and in the air, and timely inventory updates to some of your favourite department stores and grocery chains, not to mention that Solace technology also powers many of the world's leading stock exchanges and betting houses. Aside from rock solid technology, stellar customer support is one of the biggest reasons customers select Solace, and stick with them. -
26
IBM Streams
IBM
1 RatingIBM Streams analyzes a wide range of streaming data, including unstructured text, video and audio, and geospatial and sensor data. This helps organizations to spot opportunities and risks, and make decisions in real-time. -
27
Google Cloud Datastream
Google
Change data capture and replication service that is serverless and easy to use. Access to streaming data in MySQL, PostgreSQL and AlloyDB databases. BigQuery offers near-real-time analytics. Easy-to use setup with built-in security connectivity for faster time to value. A serverless platform which automatically scales without the need to provision or manage resources. Log-based mechanism reduces the load on source databases and any potential disruption. Synchronize data reliably across heterogeneous storage systems, databases, and applications with low latency while minimising impact on source performance. Easy-to-use and serverless service that scales up and down seamlessly and does not require infrastructure management will get you up and running quickly. Connect and integrate your data with the best Google Cloud services, including BigQuery, Spanner Dataflow and Data Fusion. -
28
Nussknacker
Nussknacker
0Nussknacker allows domain experts to use a visual tool that is low-code to help them create and execute real-time decisioning algorithm instead of writing code. It is used to perform real-time actions on data: real-time marketing and fraud detection, Internet of Things customer 360, Machine Learning inferring, and Internet of Things customer 360. A visual design tool for decision algorithm is an essential part of Nussknacker. It allows non-technical users, such as analysts or business people, to define decision logic in a clear, concise, and easy-to-follow manner. With a click, scenarios can be deployed for execution once they have been created. They can be modified and redeployed whenever there is a need. Nussknacker supports streaming and request-response processing modes. It uses Kafka as its primary interface in streaming mode. It supports both stateful processing and stateless processing. -
29
SAS Event Stream Processing
SAS Institute
Streaming data from operations and transactions is valuable when it is well-understood. SAS Event stream processing includes streaming data quality, analytics, and a vast array SAS and open-source machine learning and high frequency analytics for connecting to, deciphering and cleansing streaming data. It doesn't matter how fast your data moves or how many sources you pull from, all of it is under your control through a single, intuitive interface. You can create patterns and address situations from any aspect of your business, giving the ability to be agile and deal with issues as they arise. -
30
Azure Stream Analytics
Microsoft
Azure Stream Analytics is an easy-to-use, real time analytics service that's designed for mission-critical workloads. In just a few steps, you can create an end-to-end streaming pipeline that is serverless in just a few clicks. SQL--easily extensible and customizable with custom code, built-in machine learning capabilities and more advanced scenarios. You can run the most complex workloads with confidence knowing that your SLA is financially backed. -
31
Macrometa
Macrometa
We provide a geo-distributed, real-time database, stream processing, and compute runtime for event driven applications across up to 175 global edge data centers. Our platform is loved by API and app developers because it solves the most difficult problems of sharing mutable states across hundreds of locations around the world. We also have high consistency and low latency. Macrometa allows you to surgically expand your existing infrastructure to bring your application closer to your users. This allows you to improve performance and user experience, as well as comply with global data governance laws. Macrometa is a streaming, serverless NoSQL database that can be used for stream data processing, pub/sub, and compute engines. You can create stateful data infrastructure, stateful function & containers for long-running workloads, and process data streams real time. We do the ops and orchestration, you write the code. -
32
SQLstream
Guavus, a Thales company
In the field of IoT stream processing and analytics, SQLstream ranks #1 according to ABI Research. Used by Verizon, Walmart, Cisco, and Amazon, our technology powers applications on premises, in the cloud, and at the edge. SQLstream enables time-critical alerts, live dashboards, and real-time action with sub-millisecond latency. Smart cities can reroute ambulances and fire trucks or optimize traffic light timing based on real-time conditions. Security systems can detect hackers and fraudsters, shutting them down right away. AI / ML models, trained with streaming sensor data, can predict equipment failures. Thanks to SQLstream's lightning performance -- up to 13 million rows / second / CPU core -- companies have drastically reduced their footprint and cost. Our efficient, in-memory processing allows operations at the edge that would otherwise be impossible. Acquire, prepare, analyze, and act on data in any format from any source. Create pipelines in minutes not months with StreamLab, our interactive, low-code, GUI dev environment. Edit scripts instantly and view instantaneous results without compiling. Deploy with native Kubernetes support. Easy installation includes Docker, AWS, Azure, Linux, VMWare, and more -
33
Red Hat OpenShift Streams
Red Hat
Red Hat®, OpenShift®, Streams for Apache Kafka provides a simplified developer experience for building, scaling, and modernizing cloud-native apps or upgrading existing systems. Red Hat OpenShift streams for Apache Kafka make it easy to create, discover and connect to real-time data stream no matter where they're deployed. Streams are essential for delivering event-driven or data analytics applications. The seamless operation of distributed microservices, large data transfers volumes, and managed operations allow teams to focus on their strengths, speed up value, and lower operating costs. OpenShift streams for Apache Kafka is part of the Red Hat OpenShift product range, which allows you to build a wide variety of data-driven solutions. -
34
Samza
Apache Software Foundation
Samza lets you build stateful applications that can process data in real time from multiple sources, including Apache Kafka. It has been battle-tested at scale and supports flexible deployment options, including running on YARN or as a standalone program. Samza offers high throughput and low latency to instantly analyze your data. With features like host-affinity and incremental checkpoints, Samza can scale to many terabytes in state. Samza is easy-to-use with flexible deployment options YARN, Kubernetes, or standalone. The ability to run the same code to process streaming and batch data. Integrates with multiple sources, including Kafka and HDFS, AWS Kinesis Azure Eventhubs, Azure Eventhubs K-V stores, ElasticSearch, AWS Kinesis, Kafka and ElasticSearch. -
35
Apache Storm
Apache Software Foundation
Apache Storm is an open-source distributed realtime computing system that is free and open-source. Apache Storm makes it simple to process unbounded streams and data reliably, much like Hadoop did for batch processing. Apache Storm is easy to use with any programming language and is a lot fun! Apache Storm can be used for many purposes: realtime analytics and online machine learning. It can also be used with any programming language. Apache Storm is fast. A benchmark measured it at more than a million tuples per second per node. It is highly scalable, fault-tolerant and guarantees that your data will be processed. It is also easy to set up. Apache Storm can be integrated with the queueing and databases technologies you already use. Apache Storm topology processes streams of data in arbitrarily complex ways. It also partitions the streams between each stage of the computation as needed. Learn more in the tutorial. -
36
Aiven
Aiven
$200.00 per monthAiven manages your open-source data infrastructure in the cloud so that you don't have. Developers can do what is best for them: create applications. We do what we love: manage cloud data infrastructure. All solutions are open-source. You can also freely transfer data between clouds and create multi-cloud environments. You will know exactly what you will be paying and why. We combine storage, networking, and basic support costs. We will keep your Aiven software up and running. We will be there to help you if there is ever an issue. In 10 minutes, you can deploy a service on Aiven. 1. Register now - No credit card information required 2. Select your open-source service and choose the region and cloud to deploy to it 3. Select your plan and get $300 in credit 4. Click "Create service" to configure your data sources -
37
ksqlDB
Confluent
Now that your data has been in motion, it is time to make sense. Stream processing allows you to extract instant insights from your data streams but it can be difficult to set up the infrastructure. Confluent created ksqlDB to support stream processing applications. Continuously processing streams of data from your business will make your data actionable. The intuitive syntax of ksqlDB allows you to quickly access and augment Kafka data, allowing development teams to create innovative customer experiences and meet data-driven operational requirements. ksqlDB is a single solution that allows you to collect streams of data, enrich them and then serve queries on new derived streams or tables. This means that there is less infrastructure to manage, scale, secure, and deploy. You can now focus on the important things -- innovation -- with fewer moving parts in your data architecture. -
38
Aiven for Apache Kafka
Aiven
$200 per monthApache Kafka is a fully managed service that offers zero vendor lock-in, as well as all the capabilities you need to build your streaming infrastructure. You can easily set up fully managed Kafka within 10 minutes using our web console, or programmatically through our API, CLI or Terraform provider. Connect it to your existing tech stack using over 30 connectors. You can feel confident in your setup thanks to the service integrations that provide logs and metrics. Fully managed distributed data streaming platform that can be deployed in any cloud. Event-driven applications, near real-time data transfer and pipelines and stream analytics are all possible uses for this platform. Aiven's Apache Kafka is hosted and managed for you. You can create clusters, migrate clouds, upgrade versions, and deploy new nodes all with a single click. All this and more through a simple dashboard. -
39
Amazon Managed Service for Apache Flink
Amazon
$0.11 per hourAmazon Managed Service For Apache Flink is used by thousands of customers to run stream-processing applications. Amazon Managed Service Apache Flink allows you to transform and analyze streaming data using Apache Flink in real-time and integrate applications with AWS services. There are no clusters or servers to manage and no computing infrastructure to install. You only pay for the resources that you use. You can build and run Apache Flink apps without having to manage resources or clusters, or set up infrastructure. Process gigabytes per second, with latencies of subseconds and respond to events instantly. Multi-AZ deployments, APIs for lifecycle management and APIs to manage application lifecycles help you deploy highly available and durable apps. Create applications that transform data and deliver it to Amazon Simple Storage Service (Amazon S3) and Amazon OpenSearch Service. -
40
VeloDB
VeloDB
VeloDB, powered by Apache Doris is a modern database for real-time analytics at scale. In seconds, micro-batch data can be ingested using a push-based system. Storage engine with upserts, appends and pre-aggregations in real-time. Unmatched performance in real-time data service and interactive ad hoc queries. Not only structured data, but also semi-structured. Not only real-time analytics, but also batch processing. Not only run queries against internal data, but also work as an federated query engine to access external databases and data lakes. Distributed design to support linear scalability. Resource usage can be adjusted flexibly to meet workload requirements, whether on-premise or cloud deployment, separation or integration. Apache Doris is fully compatible and built on this open source software. Support MySQL functions, protocol, and SQL to allow easy integration with other tools. -
41
HarperDB
HarperDB
FreeHarperDB is an integrated distributed systems platform which combines database, caching and application functions into one technology. It allows you to deliver global back-end services at a lower cost, with higher performance and less effort. Install user-programmed apps and pre-built additions on top of data for a back end with ultra-low latencies. Distributed database with a high throughput per second, delivering orders of magnitude higher than NoSQL alternatives. Native real-time pub/sub data processing and communication via MQTT interfaces, WebSockets, and HTTP interfaces. HarperDB provides powerful data-in motion capabilities without adding additional services such as Kafka. Focus on features that will help your business grow, rather than fighting complicated infrastructure. You can't slow down the speed of light but you can reduce the amount of light between your users' data and them. -
42
Apama
Apama
Apama Streaming Analytics enables organizations to analyze and respond to IoT and fast moving data in real time, allowing them to react intelligently to events as they occur. Apama Community Edition by Software AG is a freemium version that allows users to learn about, develop, and implement streaming analytics applications. Software AG Data & Analytics Platform offers an integrated, modular and end-to-end set of world-class capabilities that are optimized for high-speed data management. It also provides connectivity and integration to all major enterprise data sources. You can choose the capabilities that you require: streaming, predictive, and visual analytics. There is also messaging and integration with other enterprise applications. You can integrate historical and other data to create models and enrich customer data. -
43
Amazon Data Firehose
Amazon
$0.075 per monthEasy to capture, transform and load streaming data. Create a stream of data, select the destination and start streaming real time data in just a few simple clicks. Automate the provisioning and scaling of compute, memory and network resources, without any ongoing administration. Transform streaming data into formats such as Apache Parquet and dynamically partition streaming without building your own pipelines. Amazon Data Firehose is the fastest way to acquire data streams, transform them, and then deliver them to data lakes, warehouses, or analytics services. Amazon Data Firehose requires you to create a stream that includes a destination, a source and the transformations required. Amazon Data Firehose continuously processes a stream, scales automatically based on data availability, and delivers the results within seconds. Select the source of your data stream, or write data with the Firehose Direct PUT (API) API. -
44
Apache Doris
The Apache Software Foundation
FreeApache Doris is an advanced data warehouse for real time analytics. It delivers lightning fast analytics on real-time, large-scale data. Ingestion of micro-batch data and streaming data within a second. Storage engine with upserts, appends and pre-aggregations in real-time. Optimize for high-concurrency, high-throughput queries using columnar storage engine, cost-based query optimizer, and vectorized execution engine. Federated querying for data lakes like Hive, Iceberg, and Hudi and databases like MySQL and PostgreSQL. Compound data types, such as Arrays, Maps and JSON. Variant data types to support auto datatype inference for JSON data. NGram bloomfilter for text search. Distributed design for linear scaling. Workload isolation, tiered storage and efficient resource management. Supports shared-nothing as well as the separation of storage from compute. -
45
SelectDB
SelectDB
$0.22 per hourSelectDB is an advanced data warehouse built on Apache Doris. It supports rapid query analysis of large-scale, real-time data. Clickhouse to Apache Doris to separate the lake warehouse, and upgrade the lake storage. Fast-hand OLAP system carries out nearly 1 billion queries every day in order to provide data services for various scenes. The original lake warehouse separation was abandoned due to problems with storage redundancy and resource seizure. Also, it was difficult to query and adjust. It was decided to use Apache Doris lakewarehouse, along with Doris's materialized views rewriting capability and automated services to achieve high-performance query and flexible governance. Write real-time data within seconds and synchronize data from databases and streams. Data storage engine with real-time update and addition, as well as real-time polymerization. -
46
Akka
Akka
Akka is a toolkit that allows you to build highly concurrent, distributed and resilient message-driven Java and Scala applications. Akka Insights, an intelligent monitoring and observability tool, is specifically designed for Akka. Actors and streams allow you to build systems that scale up by using more server resources and out by using multiple servers. You can build systems that self-heal and remain responsive to failures by using the principles of The Reactive Manifesto Akka. Distributed systems that are resilient to failure. Load balancing across nodes and adaptive routing. Cluster Sharding is used for event sourcing and CQRS. Distributed Data to ensure eventual consistency with CRDTs Backpressure and asynchronous stream processing. A great platform for microservice development is the fully async streaming HTTP server and client. Alpakka streams integrations. -
47
Kinetica
Kinetica
A cloud database that can scale to handle large streaming data sets. Kinetica harnesses modern vectorized processors to perform orders of magnitude faster for real-time spatial or temporal workloads. In real-time, track and gain intelligence from billions upon billions of moving objects. Vectorization unlocks new levels in performance for analytics on spatial or time series data at large scale. You can query and ingest simultaneously to take action on real-time events. Kinetica's lockless architecture allows for distributed ingestion, which means data is always available to be accessed as soon as it arrives. Vectorized processing allows you to do more with fewer resources. More power means simpler data structures which can be stored more efficiently, which in turn allows you to spend less time engineering your data. Vectorized processing allows for incredibly fast analytics and detailed visualizations of moving objects at large scale. -
48
Upsolver
Upsolver
Upsolver makes it easy to create a governed data lake, manage, integrate, and prepare streaming data for analysis. Only use auto-generated schema on-read SQL to create pipelines. A visual IDE that makes it easy to build pipelines. Add Upserts to data lake tables. Mix streaming and large-scale batch data. Automated schema evolution and reprocessing of previous state. Automated orchestration of pipelines (no Dags). Fully-managed execution at scale Strong consistency guarantee over object storage Nearly zero maintenance overhead for analytics-ready information. Integral hygiene for data lake tables, including columnar formats, partitioning and compaction, as well as vacuuming. Low cost, 100,000 events per second (billions every day) Continuous lock-free compaction to eliminate the "small file" problem. Parquet-based tables are ideal for quick queries. -
49
InfinyOn Cloud
InfinyOn
InfinyOn has created a programmable continuous Intelligence platform for data in motion. Infinyon Cloud, unlike other event streaming platforms built on Java is built on Rust. It delivers industry-leading scale and security for real time applications. Ready-to-use programmable connectors to shape data events in real time. Intelligent analytics pipelines can be created that automatically refine, protect, and corroborate events. To notify stakeholders and dispatch events, attach programmable connectors. Each connector can be used to either import or export data. Connectors can be deployed in one way: as a Managed Connector in which Fluvio cluster provision and manages the connector, or as a local connector in which you manually launch it as a Docker container wherever you want it. Connectors are conceptually divided into four stages, each with its own responsibilities. -
50
Insigna
Insigna
Insigna - The complete Platform for Real-time Analytics and Data Management. Insigna offers integration, automated processing, transformation, data preparation and real-time analytics to derive and deliver intelligence to various stakeholders. Insigna enables connectivity with the most popular network communication protocols, data stores, enterprise applications, and cloud platforms. Coupled with a rich set of out-of-the-box data transformation capabilities, enterprises greatly benefit from the opportunities offered by operations data generated in real-time.