Best StarTree Alternatives in 2024
Find the top alternatives to StarTree currently available. Compare ratings, reviews, pricing, and features of StarTree alternatives in 2024. Slashdot lists the best StarTree alternatives on the market that offer competing products that are similar to StarTree. Sort through StarTree alternatives below to make the best choice for your needs
-
1
ANSI SQL allows you to analyze petabytes worth of data at lightning-fast speeds with no operational overhead. Analytics at scale with 26%-34% less three-year TCO than cloud-based data warehouse alternatives. You can unleash your insights with a trusted platform that is more secure and scales with you. Multi-cloud analytics solutions that allow you to gain insights from all types of data. You can query streaming data in real-time and get the most current information about all your business processes. Machine learning is built-in and allows you to predict business outcomes quickly without having to move data. With just a few clicks, you can securely access and share the analytical insights within your organization. Easy creation of stunning dashboards and reports using popular business intelligence tools right out of the box. BigQuery's strong security, governance, and reliability controls ensure high availability and a 99.9% uptime SLA. Encrypt your data by default and with customer-managed encryption keys
-
2
Amazon Redshift
Amazon
$0.25 per hourAmazon Redshift is preferred by more customers than any other cloud data storage. Redshift powers analytic workloads for Fortune 500 companies and startups, as well as everything in between. Redshift has helped Lyft grow from a startup to multi-billion-dollar enterprises. It's easier than any other data warehouse to gain new insights from all of your data. Redshift allows you to query petabytes (or more) of structured and semi-structured information across your operational database, data warehouse, and data lake using standard SQL. Redshift allows you to save your queries to your S3 database using open formats such as Apache Parquet. This allows you to further analyze other analytics services like Amazon EMR and Amazon Athena. Redshift is the fastest cloud data warehouse in the world and it gets faster each year. The new RA3 instances can be used for performance-intensive workloads to achieve up to 3x the performance compared to any cloud data warehouse. -
3
Striim
Striim
Data integration for hybrid clouds Modern, reliable data integration across both your private cloud and public cloud. All this in real-time, with change data capture and streams. Striim was developed by the executive and technical team at GoldenGate Software. They have decades of experience in mission critical enterprise workloads. Striim can be deployed in your environment as a distributed platform or in the cloud. Your team can easily adjust the scaleability of Striim. Striim is fully secured with HIPAA compliance and GDPR compliance. Built from the ground up to support modern enterprise workloads, whether they are hosted in the cloud or on-premise. Drag and drop to create data flows among your sources and targets. Real-time SQL queries allow you to process, enrich, and analyze streaming data. -
4
Apache Doris
The Apache Software Foundation
FreeApache Doris is an advanced data warehouse for real time analytics. It delivers lightning fast analytics on real-time, large-scale data. Ingestion of micro-batch data and streaming data within a second. Storage engine with upserts, appends and pre-aggregations in real-time. Optimize for high-concurrency, high-throughput queries using columnar storage engine, cost-based query optimizer, and vectorized execution engine. Federated querying for data lakes like Hive, Iceberg, and Hudi and databases like MySQL and PostgreSQL. Compound data types, such as Arrays, Maps and JSON. Variant data types to support auto datatype inference for JSON data. NGram bloomfilter for text search. Distributed design for linear scaling. Workload isolation, tiered storage and efficient resource management. Supports shared-nothing as well as the separation of storage from compute. -
5
Rockset
Rockset
FreeReal-time analytics on raw data. Live ingest from S3, DynamoDB, DynamoDB and more. Raw data can be accessed as SQL tables. In minutes, you can create amazing data-driven apps and live dashboards. Rockset is a serverless analytics and search engine that powers real-time applications and live dashboards. You can directly work with raw data such as JSON, XML and CSV. Rockset can import data from real-time streams and data lakes, data warehouses, and databases. You can import real-time data without the need to build pipelines. Rockset syncs all new data as it arrives in your data sources, without the need to create a fixed schema. You can use familiar SQL, including filters, joins, and aggregations. Rockset automatically indexes every field in your data, making it lightning fast. Fast queries are used to power your apps, microservices and live dashboards. Scale without worrying too much about servers, shards or pagers. -
6
StarRocks
StarRocks
FreeStarRocks offers at least 300% more performance than other popular solutions, whether you're using a single or multiple tables. With a rich set connectors, you can ingest real-time data into StarRocks for the latest insights. A query engine that adapts your use cases. StarRocks allows you to scale your analytics easily without moving your data or rewriting SQL. StarRocks allows a rapid journey between data and insight. StarRocks is unmatched in performance and offers a unified OLAP system that covers the most common data analytics scenarios. StarRocks offers at least 300% faster performance than other popular solutions, whether you are working with one table or many. StarRocks' built-in memory-and-disk-based caching framework is specifically designed to minimize the I/O overhead of fetching data from external storage to accelerate query performance. -
7
VeloDB
VeloDB
VeloDB, powered by Apache Doris is a modern database for real-time analytics at scale. In seconds, micro-batch data can be ingested using a push-based system. Storage engine with upserts, appends and pre-aggregations in real-time. Unmatched performance in real-time data service and interactive ad hoc queries. Not only structured data, but also semi-structured. Not only real-time analytics, but also batch processing. Not only run queries against internal data, but also work as an federated query engine to access external databases and data lakes. Distributed design to support linear scalability. Resource usage can be adjusted flexibly to meet workload requirements, whether on-premise or cloud deployment, separation or integration. Apache Doris is fully compatible and built on this open source software. Support MySQL functions, protocol, and SQL to allow easy integration with other tools. -
8
Apache Pinot
Apache Corporation
Pinot is designed to answer OLAP questions with low latency and immutable data. Pluggable indexing technologies: Sorted Index (Bitmap Index), Inverted Index. Trino and PrestoDB are both available for querying, but joins are not currently supported. SQL-like language that supports selection and aggregation, filtering as well as group by, order, and distinct queries on data. Both an offline and a real-time table are possible. Only use real-time table to cover segments where offline data is not yet available. Customize anomaly detection flow and notification flow to detect the right anomalies. -
9
Apache Druid
Druid
Apache Druid, an open-source distributed data store, is Apache Druid. Druid's core design blends ideas from data warehouses and timeseries databases to create a high-performance real-time analytics database that can be used for a wide range of purposes. Druid combines key characteristics from each of these systems into its ingestion, storage format, querying, and core architecture. Druid compresses and stores each column separately, so it only needs to read the ones that are needed for a specific query. This allows for fast scans, ranking, groupBys, and groupBys. Druid creates indexes that are inverted for string values to allow for fast search and filter. Connectors out-of-the box for Apache Kafka and HDFS, AWS S3, stream processors, and many more. Druid intelligently divides data based upon time. Time-based queries are much faster than traditional databases. Druid automatically balances servers as you add or remove servers. Fault-tolerant architecture allows for server failures to be avoided. -
10
Arroyo
Arroyo
Scale from 0 to millions of events every second. Arroyo is shipped as a single compact binary. Run locally on MacOS, Linux or Kubernetes for development and deploy to production using Docker or Kubernetes. Arroyo is an entirely new stream processing engine that was built from the ground-up to make real time easier than batch. Arroyo has been designed so that anyone with SQL knowledge can build reliable, efficient and correct streaming pipelines. Data scientists and engineers are able to build real-time dashboards, models, and applications from end-to-end without the need for a separate streaming expert team. SQL allows you to transform, filter, aggregate and join data streams with results that are sub-second. Your streaming pipelines should not page someone because Kubernetes rescheduled your pods. Arroyo can run in a modern, elastic cloud environment, from simple container runtimes such as Fargate, to large, distributed deployments using the Kubernetes logo. -
11
Databricks Data Intelligence Platform
Databricks
The Databricks Data Intelligence Platform enables your entire organization to utilize data and AI. It is built on a lakehouse that provides an open, unified platform for all data and governance. It's powered by a Data Intelligence Engine, which understands the uniqueness in your data. Data and AI companies will win in every industry. Databricks can help you achieve your data and AI goals faster and easier. Databricks combines the benefits of a lakehouse with generative AI to power a Data Intelligence Engine which understands the unique semantics in your data. The Databricks Platform can then optimize performance and manage infrastructure according to the unique needs of your business. The Data Intelligence Engine speaks your organization's native language, making it easy to search for and discover new data. It is just like asking a colleague a question. -
12
Timeplus
Timeplus
$199 per monthTimeplus is an easy-to-use, powerful and cost-effective platform for stream processing. All in one binary, easily deployable anywhere. We help data teams in organizations of any size and industry process streaming data and historical data quickly, intuitively and efficiently. Lightweight, one binary, no dependencies. Streaming analytics and historical functionality from end-to-end. 1/10 of the cost of comparable open source frameworks Transform real-time data from the market and transactions into real-time insight. Monitor financial data using append-only streams or key-value streams. Implement real-time feature pipelines using Timeplus. All infrastructure logs, metrics and traces are consolidated on one platform. In Timeplus we support a variety of data sources through our web console UI. You can also push data using REST API or create external streams, without copying data to Timeplus. -
13
Google Cloud Bigtable
Google
Google Cloud Bigtable provides a fully managed, scalable NoSQL data service that can handle large operational and analytical workloads. Cloud Bigtable is fast and performant. It's the storage engine that grows with your data, from your first gigabyte up to a petabyte-scale for low latency applications and high-throughput data analysis. Seamless scaling and replicating: You can start with one cluster node and scale up to hundreds of nodes to support peak demand. Replication adds high availability and workload isolation to live-serving apps. Integrated and simple: Fully managed service that easily integrates with big data tools such as Dataflow, Hadoop, and Dataproc. Development teams will find it easy to get started with the support for the open-source HBase API standard. -
14
Materialize
Materialize
$0.98 per hourMaterialize is a reactive database that provides incremental view updates. Our standard SQL allows developers to easily work with streaming data. Materialize connects to many external data sources without any pre-processing. Connect directly to streaming sources such as Kafka, Postgres databases and CDC or historical data sources such as files or S3. Materialize allows you to query, join, and transform data sources in standard SQL - and presents the results as incrementally-updated Materialized views. Queries are kept current and updated as new data streams are added. With incrementally-updated views, developers can easily build data visualizations or real-time applications. It is as easy as writing a few lines SQL to build with streaming data. -
15
ClickHouse
ClickHouse
1 RatingClickHouse is an open-source OLAP database management software that is fast and easy to use. It is column-oriented, and can generate real-time analytical reports by using SQL queries. ClickHouse's performance is superior to comparable column-oriented database management software currently on the market. It processes hundreds of millions of rows to more than a million and tens if not thousands of gigabytes per second. ClickHouse makes use of all hardware available to process every query as quickly as possible. Peak processing speed for a single query is more than 2 Terabytes per Second (after decompression, only utilized columns). To reduce latency, reads in distributed setups are automatically balanced between healthy replicas. ClickHouse supports multimaster asynchronous replication, and can be deployed across multiple datacenters. Each node is equal, which prevents single points of failure. -
16
Kinetica
Kinetica
A cloud database that can scale to handle large streaming data sets. Kinetica harnesses modern vectorized processors to perform orders of magnitude faster for real-time spatial or temporal workloads. In real-time, track and gain intelligence from billions upon billions of moving objects. Vectorization unlocks new levels in performance for analytics on spatial or time series data at large scale. You can query and ingest simultaneously to take action on real-time events. Kinetica's lockless architecture allows for distributed ingestion, which means data is always available to be accessed as soon as it arrives. Vectorized processing allows you to do more with fewer resources. More power means simpler data structures which can be stored more efficiently, which in turn allows you to spend less time engineering your data. Vectorized processing allows for incredibly fast analytics and detailed visualizations of moving objects at large scale. -
17
Your cloud data platform. Access to any data you need with unlimited scalability. All your data is available to you, with the near-infinite performance and concurrency required by your organization. You can seamlessly share and consume shared data across your organization to collaborate and solve your most difficult business problems. You can increase productivity and reduce time to value by collaborating with data professionals to quickly deliver integrated data solutions from any location in your organization. Our technology partners and system integrators can help you deploy Snowflake to your success, no matter if you are moving data into Snowflake.
-
18
DoubleCloud
DoubleCloud
$0.024 per 1 GB per monthOpen source solutions that require no maintenance can save you time and money. Your engineers will enjoy working with data because it is integrated, managed and highly reliable. DoubleCloud offers a range of managed open-source services, or you can leverage the full platform's power, including data storage and visualization, orchestration, ELT and real-time visualisation. We offer leading open-source solutions like ClickHouse Kafka and Airflow with deployments on Amazon Web Services and Google Cloud. Our no-code ELT allows real-time data sync between systems. It is fast, serverless and seamlessly integrated into your existing infrastructure. Our managed open-source data visualisation allows you to visualize your data in real time by creating charts and dashboards. Our platform is designed to make engineers' lives easier. -
19
Amazon Aurora
Amazon
$0.02 per month 1 RatingAmazon Aurora is a MySQL and PostgreSQL-compatible relational database built for the cloud, that combines the performance and availability of traditional enterprise databases with the simplicity and cost-effectiveness of open source databases. Amazon Aurora is five times faster that standard MySQL databases and three time faster than standard PostgreSQL database. It offers the same security, availability, reliability, and cost-effectiveness as commercial databases, but at a fraction of the cost. Amazon Aurora is fully managed and maintained by Amazon Relational Database Service, (RDS). This automates tedious administration tasks such as hardware provisioning, database setup, patching and backups. Amazon Aurora is a distributed, fault-tolerant and self-healing storage that auto-scales up 64TB per database instance. It offers high availability and performance with up to 15 low latency read replicas, point in time recovery, continuous backup to Amazon S3, replication across threeAvailability Zones, and continuous backup to Amazon S3. -
20
WarpStream
WarpStream
$2,987 per monthWarpStream, an Apache Kafka compatible data streaming platform, is built directly on object storage. It has no inter-AZ network costs, no disks that need to be managed, and it's infinitely scalable within your VPC. WarpStream is deployed in your VPC as a stateless, auto-scaling binary agent. No local disks are required to be managed. Agents stream data directly into and out of object storage without buffering on local drives and no data tiering. Instantly create new "virtual" clusters in our control plan. Support multiple environments, teams or projects without having to manage any dedicated infrastructure. WarpStream is Apache Kafka protocol compatible, so you can continue to use your favorite tools and applications. No need to rewrite or use a proprietary SDK. Simply change the URL of your favorite Kafka library in order to start streaming. Never again will you have to choose between budget and reliability. -
21
SelectDB
SelectDB
$0.22 per hourSelectDB is an advanced data warehouse built on Apache Doris. It supports rapid query analysis of large-scale, real-time data. Clickhouse to Apache Doris to separate the lake warehouse, and upgrade the lake storage. Fast-hand OLAP system carries out nearly 1 billion queries every day in order to provide data services for various scenes. The original lake warehouse separation was abandoned due to problems with storage redundancy and resource seizure. Also, it was difficult to query and adjust. It was decided to use Apache Doris lakewarehouse, along with Doris's materialized views rewriting capability and automated services to achieve high-performance query and flexible governance. Write real-time data within seconds and synchronize data from databases and streams. Data storage engine with real-time update and addition, as well as real-time polymerization. -
22
Oxla
Oxla
$0.06 per hourOxla is a new-generation Online Analytical Process (OLAP) Database engineered for high-speed processing and efficiency. Its all-in one architecture allows rapid deployment without external dependencies and allows users to insert data and query it seamlessly. Oxla is compatible both with the PostgreSQL SQL dialect and wire protocol, making it easy to integrate with existing tools and workflows. The platform excels at both real-time processing as well as handling large, complex query, making it ideal for diverse analytical tasks. Oxla's design is optimized for modern hardware, including multi-core architectural capabilities, delivering superior performance to traditional analytical databases. It offers flexible deployment, including self hosted and cloud-based options, and provides a 1-core license that grants access to core functionality. Oxla's pay as you go pricing model ensures cost effectiveness, allowing users only to pay for the resources that they use. -
23
Imply
Imply
Imply is a real time analytics platform built on Apache Druid. It was designed to handle large scale, high performance OLAP (Online Analytical Processing). It provides real-time data ingestion and fast query performance. It also allows for complex analytical queries to be performed on massive datasets at low latency. Imply is designed for organizations who need interactive analytics, real time dashboards, and data driven decision-making. It offers a user-friendly data exploration interface, as well as advanced features like multi-tenancy and fine-grained controls for access. Imply's distributed architecture and scalability make it ideal for use cases such as streaming data analytics, real-time monitoring, and business intelligence. -
24
Presto
Presto Foundation
Presto is an open-source distributed SQL query engine that allows interactive analytic queries against any data source, from gigabytes up to petabytes. -
25
Greenplum
Greenplum Database
Greenplum Database®, an open-source data warehouse, is a fully featured, advanced, and fully functional data warehouse. It offers powerful and fast analytics on petabyte-scale data volumes. Greenplum Database is uniquely designed for big data analytics. It is powered by the most advanced cost-based query optimizer in the world, delivering high analytical query performance with large data volumes. The Apache 2 license is used to release Greenplum Database®. We would like to thank all of our community contributors. We are also open to new contributions. We encourage all contributions to the Greenplum Database community, no matter how small. Open-source, massively parallel data platform for machine learning, analytics, and AI. Rapidly create and deploy models to support complex applications in cybersecurity, predictive management, risk management, fraud detection, among other areas. The fully integrated, open-source analytics platform is now available. -
26
DeltaStream
DeltaStream
DeltaStream is an integrated serverless streaming processing platform that integrates seamlessly with streaming storage services. Imagine it as a compute layer on top your streaming storage. It offers streaming databases and streaming analytics along with other features to provide an integrated platform for managing, processing, securing and sharing streaming data. DeltaStream has a SQL-based interface that allows you to easily create stream processing apps such as streaming pipelines. It uses Apache Flink, a pluggable stream processing engine. DeltaStream is much more than a query-processing layer on top Kafka or Kinesis. It brings relational databases concepts to the world of data streaming, including namespacing, role-based access control, and enables you to securely access and process your streaming data, regardless of where it is stored. -
27
Apache Spark
Apache Software Foundation
Apache Spark™, a unified analytics engine that can handle large-scale data processing, is available. Apache Spark delivers high performance for streaming and batch data. It uses a state of the art DAG scheduler, query optimizer, as well as a physical execution engine. Spark has over 80 high-level operators, making it easy to create parallel apps. You can also use it interactively via the Scala, Python and R SQL shells. Spark powers a number of libraries, including SQL and DataFrames and MLlib for machine-learning, GraphX and Spark Streaming. These libraries can be combined seamlessly in one application. Spark can run on Hadoop, Apache Mesos and Kubernetes. It can also be used standalone or in the cloud. It can access a variety of data sources. Spark can be run in standalone cluster mode on EC2, Hadoop YARN and Mesos. Access data in HDFS and Alluxio. -
28
Streaming service is a streaming service that allows developers and data scientists to stream real-time events. It is serverless and Apache Kafka compatible. Streaming can be integrated with Oracle Cloud Infrastructure, Database, GoldenGate, Integration Cloud, and Oracle Cloud Infrastructure (OCI). The service provides integrations for hundreds third-party products, including databases, big data, DevOps, and SaaS applications. Data engineers can easily create and manage big data pipelines. Oracle manages all infrastructure and platform management, including provisioning, scaling and security patching. Streaming can provide state management to thousands of consumers with the help of consumer groups. This allows developers to easily create applications on a large scale.
-
29
IBM Streams
IBM
1 RatingIBM Streams analyzes a wide range of streaming data, including unstructured text, video and audio, and geospatial and sensor data. This helps organizations to spot opportunities and risks, and make decisions in real-time. -
30
Confluent
Confluent
Apache Kafka®, with Confluent, has an infinite retention. Be infrastructure-enabled, not infrastructure-restricted Legacy technologies require you to choose between being real-time or highly-scalable. Event streaming allows you to innovate and win by being both highly-scalable and real-time. Ever wonder how your rideshare app analyses massive amounts of data from multiple sources in order to calculate real-time ETA. Wondering how your credit card company analyzes credit card transactions from all over the world and sends fraud notifications in real time? Event streaming is the answer. Microservices are the future. A persistent bridge to the cloud can enable your hybrid strategy. Break down silos to demonstrate compliance. Gain real-time, persistent event transport. There are many other options. -
31
Google Cloud Dataflow
Google
Unified stream and batch data processing that is serverless, fast, cost-effective, and low-cost. Fully managed data processing service. Automated provisioning of and management of processing resource. Horizontal autoscaling worker resources to maximize resource use Apache Beam SDK is an open-source platform for community-driven innovation. Reliable, consistent processing that works exactly once. Streaming data analytics at lightning speed Dataflow allows for faster, simpler streaming data pipeline development and lower data latency. Dataflow's serverless approach eliminates the operational overhead associated with data engineering workloads. Dataflow allows teams to concentrate on programming and not managing server clusters. Dataflow's serverless approach eliminates operational overhead from data engineering workloads, allowing teams to concentrate on programming and not managing server clusters. Dataflow automates provisioning, management, and utilization of processing resources to minimize latency. -
32
SQLstream
Guavus, a Thales company
In the field of IoT stream processing and analytics, SQLstream ranks #1 according to ABI Research. Used by Verizon, Walmart, Cisco, and Amazon, our technology powers applications on premises, in the cloud, and at the edge. SQLstream enables time-critical alerts, live dashboards, and real-time action with sub-millisecond latency. Smart cities can reroute ambulances and fire trucks or optimize traffic light timing based on real-time conditions. Security systems can detect hackers and fraudsters, shutting them down right away. AI / ML models, trained with streaming sensor data, can predict equipment failures. Thanks to SQLstream's lightning performance -- up to 13 million rows / second / CPU core -- companies have drastically reduced their footprint and cost. Our efficient, in-memory processing allows operations at the edge that would otherwise be impossible. Acquire, prepare, analyze, and act on data in any format from any source. Create pipelines in minutes not months with StreamLab, our interactive, low-code, GUI dev environment. Edit scripts instantly and view instantaneous results without compiling. Deploy with native Kubernetes support. Easy installation includes Docker, AWS, Azure, Linux, VMWare, and more -
33
Redpanda
Redpanda Data
You can deliver customer experiences like never before with breakthrough data streaming capabilities Both the ecosystem and Kafka API are compatible. Redpanda BulletPredictable low latency with zero data loss. Redpanda BulletUp to 10x faster than Kafka Redpanda BulletEnterprise-grade support and hotfixes. Redpanda BulletAutomated backups for S3/GCS. Redpanda Bullet100% freedom of routine Kafka operations. Redpanda BulletSupports for AWS/GCP. Redpanda was built from the ground up to be easy to install and get running quickly. Redpanda's power will be evident once you have tried it in production. You can use the more advanced Redpanda functions. We manage all aspects of provisioning, monitoring, as well as upgrades. We do not have access to your cloud credentials. Sensitive data never leaves your environment. You can have it provisioned, operated, maintained, and updated for you. Configurable instance types. As your needs change, you can expand the cluster. -
34
ScyllaDB
ScyllaDB
The fastest NoSQL database in the world. The fastest NoSQL database available, capable of millions IOPS per node with less than 1 millisecond latency. This database will accelerate your application performance. Scylla, a drop-in Apache Cassandra and Amazon DynamoDB alternative, powers your applications with extreme throughput and ultra-low latency. To power modern, high-performance applications, we used the best features of high availability databases to create a NoSQL database that is significantly more efficient, fault-tolerant, and resource-efficient. This high-availability database is built from scratch in C++ for Linux. Scylla unleashes your infrastructure's true potential for running high-throughput/low-latency workloads. -
35
Amazon Timestream
Amazon
Amazon Timestream is a fast, scalable and serverless time series data service for IoT/operational applications. It makes it possible to store and analyze trillions per day up to 1000 times faster than traditional relational databases and at as low as 1/10th of the cost. Amazon Timestream helps you save time and money when managing the lifecycles of time series data. It stores recent data in memory and moves historical data to a cost-optimized storage tier according to user defined policies. Amazon Timestream's purpose-built query tool allows you to access and analyze both recent and historic data simultaneously, without having to specify in the query whether the data is in the in-memory tier or the cost-optimized. Amazon Timestream's built-in time series analytics functions allow you to identify trends and patterns within your data in real-time. -
36
Aerospike
Aerospike
Aerospike is the global leader for next-generation, real time NoSQL data solutions at any scale. Aerospike helps enterprises overcome seemingly impossible data bottlenecks and compete with other companies at a fraction of the cost and complexity of legacy NoSQL databases. Aerospike's Hybrid Memory Architecture™ is a patented technology that unlocks the full potential of modern hardware and delivers previously unimaginable value. It does this by delivering unimaginable value from huge amounts of data at both the edge, core, and in the cloud. Aerospike empowers customers with the ability to instantly combat fraud, dramatically increase shopping cart sizes, deploy global digital payment networks, and provide instant, one-to-1 personalization for millions. Aerospike customers include Airtel and Banca d'Italia as well as Snap, Verizon Media, Wayfair, PayPal, Snap, Verizon Media, and Nielsen. The company's headquarters is in Mountain View, California. Additional locations are in London, Bengaluru, India, and Tel Aviv in Israel. -
37
Aiven
Aiven
$200.00 per monthAiven manages your open-source data infrastructure in the cloud so that you don't have. Developers can do what is best for them: create applications. We do what we love: manage cloud data infrastructure. All solutions are open-source. You can also freely transfer data between clouds and create multi-cloud environments. You will know exactly what you will be paying and why. We combine storage, networking, and basic support costs. We will keep your Aiven software up and running. We will be there to help you if there is ever an issue. In 10 minutes, you can deploy a service on Aiven. 1. Register now - No credit card information required 2. Select your open-source service and choose the region and cloud to deploy to it 3. Select your plan and get $300 in credit 4. Click "Create service" to configure your data sources -
38
Trino
Trino
FreeTrino is an engine that runs at incredible speeds. Fast-distributed SQL engine for big data analytics. Helps you explore the data universe. Trino is an extremely parallel and distributed query-engine, which is built from scratch for efficient, low latency analytics. Trino is used by the largest organizations to query data lakes with exabytes of data and massive data warehouses. Supports a wide range of use cases including interactive ad-hoc analysis, large batch queries that take hours to complete, and high volume apps that execute sub-second queries. Trino is a ANSI SQL query engine that works with BI Tools such as R Tableau Power BI Superset and many others. You can natively search data in Hadoop S3, Cassandra MySQL and many other systems without having to use complex, slow and error-prone copying processes. Access data from multiple systems in a single query. -
39
Vertica
OpenText
The Unified Analytics Warehouse. The Unified Analytics Warehouse is the best place to find high-performing analytics and machine learning at large scale. Tech research analysts are seeing new leaders as they strive to deliver game-changing big data analytics. Vertica empowers data-driven companies so they can make the most of their analytics initiatives. It offers advanced time-series, geospatial, and machine learning capabilities, as well as data lake integration, user-definable extensions, cloud-optimized architecture and more. Vertica's Under the Hood webcast series allows you to dive into the features of Vertica - delivered by Vertica engineers, technical experts, and others - and discover what makes it the most scalable and scalable advanced analytical data database on the market. Vertica supports the most data-driven disruptors around the globe in their pursuit for industry and business transformation. -
40
ksqlDB
Confluent
Now that your data has been in motion, it is time to make sense. Stream processing allows you to extract instant insights from your data streams but it can be difficult to set up the infrastructure. Confluent created ksqlDB to support stream processing applications. Continuously processing streams of data from your business will make your data actionable. The intuitive syntax of ksqlDB allows you to quickly access and augment Kafka data, allowing development teams to create innovative customer experiences and meet data-driven operational requirements. ksqlDB is a single solution that allows you to collect streams of data, enrich them and then serve queries on new derived streams or tables. This means that there is less infrastructure to manage, scale, secure, and deploy. You can now focus on the important things -- innovation -- with fewer moving parts in your data architecture. -
41
InfluxDB
InfluxData
$0InfluxDB is a purpose-built data platform designed to handle all time series data, from users, sensors, applications and infrastructure — seamlessly collecting, storing, visualizing, and turning insight into action. With a library of more than 250 open source Telegraf plugins, importing and monitoring data from any system is easy. InfluxDB empowers developers to build transformative IoT, monitoring and analytics services and applications. InfluxDB’s flexible architecture fits any implementation — whether in the cloud, at the edge or on-premises — and its versatility, accessibility and supporting tools (client libraries, APIs, etc.) make it easy for developers at any level to quickly build applications and services with time series data. Optimized for developer efficiency and productivity, the InfluxDB platform gives builders time to focus on the features and functionalities that give their internal projects value and their applications a competitive edge. To get started, InfluxData offers free training through InfluxDB University. -
42
Apache Flink
Apache Software Foundation
Apache Flink is a distributed processing engine and framework for stateful computations using unbounded and bounded data streams. Flink can be used in all cluster environments and perform computations at any scale and in-memory speed. A stream of events can be used to produce any type of data. All data, including credit card transactions, machine logs, sensor measurements, and user interactions on a website, mobile app, are generated as streams. Apache Flink excels in processing both unbounded and bound data sets. Flink's runtime can run any type of application on unbounded stream streams thanks to its precise control of state and time. Bounded streams are internal processed by algorithms and data structure that are specifically designed to process fixed-sized data sets. This results in excellent performance. Flink can be used with all of the resource managers previously mentioned. -
43
Informatica Data Engineering Streaming
Informatica
AI-powered Informatica Data Engineering streaming allows data engineers to ingest and process real-time streaming data in order to gain actionable insights. -
44
Amazon Kinesis
Amazon
You can quickly collect, process, analyze, and analyze video and data streams. Amazon Kinesis makes it easy for you to quickly and easily collect, process, analyze, and interpret streaming data. Amazon Kinesis provides key capabilities to process streaming data at any scale cost-effectively, as well as the flexibility to select the tools that best fit your application's requirements. Amazon Kinesis allows you to ingest real-time data, including video, audio, website clickstreams, application logs, and IoT data for machine learning, analytics, or other purposes. Amazon Kinesis allows you to instantly process and analyze data, rather than waiting for all the data to be collected before processing can begin. Amazon Kinesis allows you to ingest buffer and process streaming data instantly, so you can get insights in seconds or minutes, instead of waiting for hours or days. -
45
Azure Stream Analytics
Microsoft
Azure Stream Analytics is an easy-to-use, real time analytics service that's designed for mission-critical workloads. In just a few steps, you can create an end-to-end streaming pipeline that is serverless in just a few clicks. SQL--easily extensible and customizable with custom code, built-in machine learning capabilities and more advanced scenarios. You can run the most complex workloads with confidence knowing that your SLA is financially backed. -
46
DataStax
DataStax
The Open, Multi-Cloud Stack to Modern Data Apps. Built on Apache Cassandra™, an open-source Apache Cassandra™. Global scale and 100% uptime without vendor lock in You can deploy on multi-clouds, open-source, on-prem and Kubernetes. For a lower TCO, use elastic and pay-as you-go. Stargate APIs allow you to build faster with NoSQL, reactive, JSON and REST. Avoid the complexity of multiple OSS projects or APIs that don’t scale. It is ideal for commerce, mobile and AI/ML. Get building modern data applications with Astra, a database-as-a-service powered by Apache Cassandra™. Richly interactive apps that are viral-ready and elastic using REST, GraphQL and JSON. Pay-as you-go Apache Cassandra DBaaS which scales easily and affordably -
47
Cloudera DataFlow
Cloudera
You can manage your data from the edge to the cloud with a simple, no-code approach to creating sophisticated streaming applications. -
48
Oracle Exadata
Oracle
Oracle Exadata is the best location to run Oracle Database. It simplifies digital transformations, increases database performance, and reduces costs. According to Wikibon's analysis, Oracle Exadata provides customers with greater availability, higher performance, and upto 40% lower costs. Customers can modernize their database infrastructure, move enterprise apps to the cloud, and quickly implement digital transformations with Oracle Cloud Infrastructure, Oracle Cloud@Customer and on-premises deployment options. Customers can run Oracle Database with Oracle Exadata. Customers can easily move workloads between their on-premises data centers, Cloud@Customer deployments and Oracle Cloud Infrastructure. This allows them to modernize their operations and reduce costs. -
49
CockroachDB
Cockroach Labs
1 RatingCockroachDB: Cloud-native distributed SQL. Your cloud applications deserve a cloud-native database. Cloud-based apps and services need a database that can scale across clouds, reduces operational complexity, and improves reliability. CockroachDB provides resilient, distributed SQL with ACID transactions. Data partitioned by geography is also available. Combining CockroachDB and orchestration tools such as Mesosphere DC/OS and Kubernetes to automate mission-critical applications can speed up operations. -
50
SAP HANA
SAP
SAP HANA is an in-memory database with high performance that accelerates data-driven decision-making and actions. It supports all workloads and provides the most advanced analytics on multi-model data on premise and in cloud.