Best AWS Glue Alternatives in 2024

Find the top alternatives to AWS Glue currently available. Compare ratings, reviews, pricing, and features of AWS Glue alternatives in 2024. Slashdot lists the best AWS Glue alternatives on the market that offer competing products that are similar to AWS Glue. Sort through AWS Glue alternatives below to make the best choice for your needs

  • 1
    Google Cloud BigQuery Reviews
    See Software
    Learn More
    Compare Both
    ANSI SQL allows you to analyze petabytes worth of data at lightning-fast speeds with no operational overhead. Analytics at scale with 26%-34% less three-year TCO than cloud-based data warehouse alternatives. You can unleash your insights with a trusted platform that is more secure and scales with you. Multi-cloud analytics solutions that allow you to gain insights from all types of data. You can query streaming data in real-time and get the most current information about all your business processes. Machine learning is built-in and allows you to predict business outcomes quickly without having to move data. With just a few clicks, you can securely access and share the analytical insights within your organization. Easy creation of stunning dashboards and reports using popular business intelligence tools right out of the box. BigQuery's strong security, governance, and reliability controls ensure high availability and a 99.9% uptime SLA. Encrypt your data by default and with customer-managed encryption keys
  • 2
    TiMi Reviews
    See Software
    Learn More
    Compare Both
    TIMi allows companies to use their corporate data to generate new ideas and make crucial business decisions more quickly and easily than ever before. The heart of TIMi’s Integrated Platform. TIMi's ultimate real time AUTO-ML engine. 3D VR segmentation, visualization. Unlimited self service business Intelligence. TIMi is a faster solution than any other to perform the 2 most critical analytical tasks: data cleaning, feature engineering, creation KPIs, and predictive modeling. TIMi is an ethical solution. There is no lock-in, just excellence. We guarantee you work in complete serenity, without unexpected costs. TIMi's unique software infrastructure allows for maximum flexibility during the exploration phase, and high reliability during the production phase. TIMi allows your analysts to test even the most crazy ideas.
  • 3
    Minitab Connect Reviews
    The most accurate, complete, and timely data provides the best insight. Minitab Connect empowers data users across the enterprise with self service tools to transform diverse data into a network of data pipelines that feed analytics initiatives, foster collaboration and foster organizational-wide collaboration. Users can seamlessly combine and explore data from various sources, including databases, on-premise and cloud apps, unstructured data and spreadsheets. Automated workflows make data integration faster and provide powerful data preparation tools that allow for transformative insights. Data integration tools that are intuitive and flexible allow users to connect and blend data from multiple sources such as data warehouses, IoT devices and cloud storage.
  • 4
    Pentaho Reviews
    Illuminate dark data and accelerate data-driven transformation with intelligent data operations to enable an edge-to-cloud data fabric. Pentaho products automate onboarding, integrating, governing, and publishing trusted data, with an intelligent composable data platform to automate data management needs.
  • 5
    Composable DataOps Platform Reviews
    Composable is an enterprise-grade DataOps platform designed for business users who want to build data-driven products and create data intelligence solutions. It can be used to design data-driven products that leverage disparate data sources, live streams, and event data, regardless of their format or structure. Composable offers a user-friendly, intuitive dataflow visual editor, built-in services that facilitate data engineering, as well as a composable architecture which allows abstraction and integration of any analytical or software approach. It is the best integrated development environment for discovering, managing, transforming, and analysing enterprise data.
  • 6
    IRI Voracity Reviews

    IRI Voracity

    IRI, The CoSort Company

    IRI Voracity is an end-to-end software platform for fast, affordable, and ergonomic data lifecycle management. Voracity speeds, consolidates, and often combines the key activities of data discovery, integration, migration, governance, and analytics in a single pane of glass, built on Eclipse™. Through its revolutionary convergence of capability and its wide range of job design and runtime options, Voracity bends the multi-tool cost, difficulty, and risk curves away from megavendor ETL packages, disjointed Apache projects, and specialized software. Voracity uniquely delivers the ability to perform data: * profiling and classification * searching and risk-scoring * integration and federation * migration and replication * cleansing and enrichment * validation and unification * masking and encryption * reporting and wrangling * subsetting and testing Voracity runs on-premise, or in the cloud, on physical or virtual machines, and its runtimes can also be containerized or called from real-time applications or batch jobs.
  • 7
    Azure Data Catalog Reviews

    Azure Data Catalog

    Microsoft

    $1 per user per month
    The new world of data means that you can spend more time looking at data than you are analyzing it. Azure Data Catalog is an enterprise-wide data catalog that makes data asset discovery easy. It is a fully managed service that allows you, from analyst to data scientist, to use it to register, enrich, discover and understand data sources. You can work with data using the tool you prefer. Data Catalog allows you to find the data you need, and you can use it in any tool you choose. Data Catalog ensures that your data is always available. It makes it easy to find and use the data you need, and provides an intuitive user interface. Increase adoption and continuous value generation across your data ecosystem. Data Catalog provides tips, tricks, as well as unwritten rules that can be used to create an environment where everyone can benefit. Data Catalog makes it possible for everyone to contribute. Democratize data asset discovery.
  • 8
    HERE Reviews

    HERE

    HERE Technologies

    $0.08 per GB
    1 Rating
    HERE is the #1 Location Platform For Developers*, ranked above Google Mapbox and TomTom in terms of mapping quality. Switch to a higher-quality offering and take advantage more monetization opportunities. Rich location data, intelligent products, and powerful tools can all be combined to propel your business forward. HERE allows you to add location-aware capabilities into your apps and online services. You get free access to over 20 market-leading AAPs, including Mapping and Geocoding, Routing and Traffic, Weather, and many more. Sign up for HERE Freemium to get access to the HERE XYZ Map Builder, which offers 5GB of storage for all your geodata. No matter what your level of skill, you can get started with industry-leading mapping technology and location technology. *Counterpoint 2019 Report
  • 9
    Airbyte Reviews

    Airbyte

    Airbyte

    $2.50 per credit
    All your ELT data pipelines, including custom ones, will be up and running in minutes. Your team can focus on innovation and insights. Unify all your data integration pipelines with one open-source ELT platform. Airbyte can meet all the connector needs of your data team, no matter how complex or large they may be. Airbyte is a data integration platform that scales to meet your high-volume or custom needs. From large databases to the long tail API sources. Airbyte offers a long list of connectors with high quality that can adapt to API and schema changes. It is possible to unify all native and custom ELT. Our connector development kit allows you to quickly edit and create new connectors from pre-built open-source ones. Transparent and scalable pricing. Finally, transparent and predictable pricing that scales with data needs. No need to worry about volume. No need to create custom systems for your internal scripts or database replication.
  • 10
    Azure Data Factory Reviews
    Accelerate data integration Azure Data Factory is a service that integrates data silos. It is designed for all levels of data integration. You can easily create ETL and ELT processes in the intuitive visual environment. Or, you can write your own code. Visually integrate data sources with over 90+ pre-built and maintenance-free connectors. The serverless integration service takes care of the rest.
  • 11
    AWS Data Pipeline Reviews
    AWS Data Pipeline, a web service, allows you to reliably process and transfer data between different AWS compute- and storage services as well as on premises data sources at specific intervals. AWS Data Pipeline allows you to access your data wherever it is stored, transform it and process it at scale, then transfer it to AWS services like Amazon S3, Amazon RDS and Amazon DynamoDB. AWS Data Pipeline makes it easy to create complex data processing workloads that can be fault-tolerant, repeatable, high-availability, and reliable. You don't need to worry about resource availability, managing intertask dependencies, retrying transient errors or timeouts in individual task, or creating a fail notification system. AWS Data Pipeline allows you to move and process data previously stored in on-premises silos.
  • 12
    Alation Reviews
    What if your data had a recommendation engine? Automated data inventory was created. A searchable catalog showed user behavior. Smart recommendations were made inline by the system as you typed queries. Alation, the first enterprise-wide collaborative data catalog, makes all this possible. It's a powerful tool that dramatically increases the productivity of analysts and the accuracy of analytics. It also empowers business decision-making for everyone. Alation provides proactive recommendations to data users through applications. Google inspired us to create a simple interface that connects the language of your business with the technical schema of your data. No more is it difficult to find the data you need due to complicated semantic translations. Are you unfamiliar with the data environment and unsure which data to use in your query. Alation allows you to build your query and provides inline recommendations that indicate whether data is trustworthy.
  • 13
    Apache Atlas Reviews

    Apache Atlas

    Apache Software Foundation

    Atlas is a flexible and extensible set core foundational governance services that enable enterprises to efficiently and effectively meet their compliance requirements within Hadoop. It also allows integration with the entire enterprise data ecosystem. Apache Atlas offers open metadata management and governance capabilities that allow organizations to create a catalog of their data assets, classify, govern and provide collaboration capabilities around these assets for data scientists, analysts, and the data governance group. Pre-defined types to manage various Hadoop and non Hadoop metadata. Ability to create new types to manage metadata. Types can inherit from other types, and can have simple attributes, complex attributes, and object references. Type instances, also known as entities, are able to capture metadata object details and their relationships. REST APIs allow for easier integration with types and instances.
  • 14
    Google Cloud Dataflow Reviews
    Unified stream and batch data processing that is serverless, fast, cost-effective, and low-cost. Fully managed data processing service. Automated provisioning of and management of processing resource. Horizontal autoscaling worker resources to maximize resource use Apache Beam SDK is an open-source platform for community-driven innovation. Reliable, consistent processing that works exactly once. Streaming data analytics at lightning speed Dataflow allows for faster, simpler streaming data pipeline development and lower data latency. Dataflow's serverless approach eliminates the operational overhead associated with data engineering workloads. Dataflow allows teams to concentrate on programming and not managing server clusters. Dataflow's serverless approach eliminates operational overhead from data engineering workloads, allowing teams to concentrate on programming and not managing server clusters. Dataflow automates provisioning, management, and utilization of processing resources to minimize latency.
  • 15
    IBM DataStage Reviews
    Cloud-native data integration with IBM Cloud Pak data enables you to accelerate AI innovation AI-powered data integration from anywhere. Your AI and analytics can only be as good as the data they are powered by. IBM®, DataStage®, for IBM Cloud Pak®, for Data provides high-quality data through a container-based architecture. It combines industry-leading data integration, DataOps, governance, and analytics on one data and AI platform. Automation speeds up administrative tasks, helping to reduce TCO. AI-based design accelerators, out-of-the box integration with DataOps or data science services accelerate AI innovation. Multicloud integration and parallelism allow you to deliver trusted data across hybrid and multicloud environments. The IBM Cloud Pak for Data platform allows you to manage the data and analytics lifecycle. Data science, event messaging, and data warehousing are some of the services offered. Automated load balancing and parallel engine.
  • 16
    Apache Beam Reviews

    Apache Beam

    Apache Software Foundation

    This is the easiest way to perform batch and streaming data processing. For mission-critical production workloads, write once and run anywhere data processing. Beam can read your data from any supported source, whether it's on-prem and in the cloud. Beam executes your business logic in both batch and streaming scenarios. Beam converts the results of your data processing logic into the most popular data sinks. A single programming model that can be used for both streaming and batch use cases. This is a simplified version of the code for all members of your data and applications teams. Apache Beam is extensible. TensorFlow Extended, Apache Hop and other projects built on Apache Beam are examples of Apache Beam's extensibility. Execute pipelines in multiple execution environments (runners), allowing flexibility and avoiding lock-in. Open, community-based development and support are available to help you develop your application and meet your specific needs.
  • 17
    Google Cloud Data Catalog Reviews
    Fully managed and highly scalable metadata and data discovery service. New customers receive $300 in Google Cloud credits for free during the Free Trial. All customers receive up to 1 MiB business or ingested meta data storage and 1,000,000 API calls free of charge. A simple, but powerful faceted search interface allows you to pinpoint your data. Automatically sync technical metadata and create schematized tags to support business metadata. Cloud Data Loss Prevention integration allows you to automatically tag sensitive data. Access your data immediately and scale without the need to manage or set up infrastructure. With a powerful UI built with the same search technology that Gmail or API access, empower any member of the team to find and tag data. Data Catalog is fully managed so that you can easily start and scale. Cloud IAM integrations and Cloud DLP integrations allow you to enforce data security policies and ensure compliance.
  • 18
    Hevo Reviews
    Hevo Data is a no-code, bi-directional data pipeline platform specially built for modern ETL, ELT, and Reverse ETL Needs. It helps data teams streamline and automate org-wide data flows that result in a saving of ~10 hours of engineering time/week and 10x faster reporting, analytics, and decision making. The platform supports 100+ ready-to-use integrations across Databases, SaaS Applications, Cloud Storage, SDKs, and Streaming Services. Over 500 data-driven companies spread across 35+ countries trust Hevo for their data integration needs.
  • 19
    Matillion Reviews
    Cloud-Native ETL tool. You can load and transform data to your cloud data warehouse in minutes. We have redesigned the traditional ETL process to create a solution for data integration in the cloud. Our solution makes use of the cloud's near-infinite storage capacity, which means that your projects have near-infinite scaling. We reduce the complexity of moving large amounts data by working in the cloud. In just fifteen minutes, you can process a billion rows and go live in five minutes. Modern businesses need to harness their data to gain greater business insight. Matillion can help you take your data journey to the next level by migrating, extracting, and transforming your data in cloud. This will allow you to gain new insights as well as make better business decisions.
  • 20
    Informatica Intelligent Data Management Cloud Reviews
    Our AI-powered Intelligent Data Platform, which is modular and comprehensive, is the best in the industry. It allows you to unlock the potential of data in your enterprise and empowers you with the ability to solve complex problems. Our platform sets a new standard in enterprise-class data management. We offer best-in-class products, and an integrated platform that unifies them. This allows you to power your business with intelligent information. You can connect to any data source, and scale with confidence. A global platform processes more than 15 trillion cloud transactions each month. A global platform that delivers trusted data at scale across all data management use cases will help you future-proof your business. Our AI-powered architecture supports integration patterns, allowing you to grow and develop at your own pace. Our solution is modular and API-driven.
  • 21
    Stitch Reviews
    Stitch is a cloud-based platform that allows you to extract, transform, load data. Stitch is used by more than 1000 companies to move billions records daily from SaaS databases and applications into data warehouses or data lakes.
  • 22
    Semarchy xDI Reviews
    Semarchy's flexible, unified data platform will help you make better business decisions across your organization. xDI is the high-performance, flexible, extensible data integration that integrates all your data for all types and uses. Its single technology can federate all forms of data integration and maps business rules into executable code. xDI supports multi-cloud environments, on-premise, hybrid, and cloud environments.
  • 23
    Xplenty Reviews

    Xplenty

    Xplenty Data Integration

    Xplenty is a scalable data delivery and integration software that allows large businesses and SMBs to prepare and transfer data to the cloud for analytics. Xplenty features include data transformations and drag-and-drop interface. It also integrates with over 100 data storages and SaaS apps. Developers can easily add Xplenty to their data solution stack. Xplenty allows users to schedule jobs, track job progress, and keep track of job status.
  • 24
    StreamSets Reviews

    StreamSets

    StreamSets

    $1000 per month
    StreamSets DataOps Platform. An end-to-end data integration platform to build, run, monitor and manage smart data pipelines that deliver continuous data for DataOps.
  • 25
    Denodo Reviews

    Denodo

    Denodo Technologies

    The core technology that enables modern data integration and data management. Connect disparate, structured and unstructured data sources quickly. Catalog your entire data ecosystem. The data is kept in the source and can be accessed whenever needed. Adapt data models to the consumer's needs, even if they come from multiple sources. Your back-end technologies can be hidden from end users. You can secure the virtual model and use it to consume standard SQL and other formats such as SOAP, REST, SOAP, and OData. Access to all types data is easy. Data integration and data modeling capabilities are available. Active Data Catalog and self service capabilities for data and metadata discovery and preparation. Full data security and governance capabilities. Data queries executed quickly and intelligently. Real-time data delivery in all formats. Data marketplaces can be created. Data-driven strategies can be made easier by separating business applications and data systems.
  • 26
    Data Virtuality Reviews
    Connect and centralize data. Transform your data landscape into a flexible powerhouse. Data Virtuality is a data integration platform that allows for instant data access, data centralization, and data governance. Logical Data Warehouse combines materialization and virtualization to provide the best performance. For high data quality, governance, and speed-to-market, create your single source data truth by adding a virtual layer to your existing data environment. Hosted on-premises or in the cloud. Data Virtuality offers three modules: Pipes Professional, Pipes Professional, or Logical Data Warehouse. You can cut down on development time up to 80% Access any data in seconds and automate data workflows with SQL. Rapid BI Prototyping allows for a significantly faster time to market. Data quality is essential for consistent, accurate, and complete data. Metadata repositories can be used to improve master data management.
  • 27
    TIBCO Cloud Metadata Reviews
    The lack of integration between the different metadata used in IT, operations and compliance is a major problem in metadata management. TIBCO Cloud™, Metadata software provides a single solution that covers all your metadata, including data dictionaries and business glossaries. Built-in machine learning (ML), and artificial intelligence (AI), algorithms allow for metadata classification and data linesages (horizontal vertical, regulatory). You need to provide the data context, coherency and control that you need to achieve the best efficiency, best performance, smartest decision-making, and highest level of collaboration across all your departments and teams. For effective execution, analysis, governance, and governance, you need consistent metadata about your operations, compliance, and analytics. Instead of using multiple solutions, choose one. You can discover, harvest, and manage metadata for all your applications, data lakes, enterprise warehouses, data lakes, APIs, streaming sources, and social media.
  • 28
    Talend Data Catalog Reviews
    Talend Data Catalog provides your organization with a single point of control for all your data. Data Catalog provides robust tools for search, discovery, and connectors that allow you to extract metadata from almost any data source. It makes it easy to manage your data pipelines, protect your data, and accelerate your ETL process. Data Catalog automatically crawls, profiles and links all your metadata. Data Catalog automatically documents up to 80% of the data associated with it. Smart relationships and machine learning keep the data current and up-to-date, ensuring that the user has the most recent data. Data governance can be made a team sport by providing a single point of control that allows you to collaborate to improve data accessibility and accuracy. With intelligent data lineage tracking and compliance tracking, you can support data privacy and regulatory compliance.
  • 29
    Lyftrondata Reviews
    Lyftrondata can help you build a governed lake, data warehouse or migrate from your old database to a modern cloud-based data warehouse. Lyftrondata makes it easy to create and manage all your data workloads from one platform. This includes automatically building your warehouse and pipeline. It's easy to share the data with ANSI SQL, BI/ML and analyze it instantly. You can increase the productivity of your data professionals while reducing your time to value. All data sets can be defined, categorized, and found in one place. These data sets can be shared with experts without coding and used to drive data-driven insights. This data sharing capability is ideal for companies who want to store their data once and share it with others. You can define a dataset, apply SQL transformations, or simply migrate your SQL data processing logic into any cloud data warehouse.
  • 30
    IBM InfoSphere Information Server Reviews
    Cloud environments can be quickly set up for quick development, testing, and productivity for your IT staff and business users. Comprehensive data governance for business users will reduce the risks and cost of maintaining your data lakes. You can save money by providing consistent, timely, and clean information for your data lakes, big data projects, and data warehouses. Also, consolidate applications and retire outdated databases. Automatic schema propagation can be used to accelerate job generation, type-ahead searching, and backwards capabilities. All this while designing once and executing everywhere. With a cognitive design that recognizes patterns and suggests ways to use them, you can create data integration flows and enforce quality rules and data governance. You can improve visibility and information governance by creating authoritative views of information that are complete and authoritative.
  • 31
    K2View Reviews
    K2View believes that every enterprise should be able to leverage its data to become as disruptive and agile as possible. We enable this through our Data Product Platform, which creates and manages a trusted dataset for every business entity – on demand, in real time. The dataset is always in sync with its sources, adapts to changes on the fly, and is instantly accessible to any authorized data consumer. We fuel operational use cases, including customer 360, data masking, test data management, data migration, and legacy application modernization – to deliver business outcomes at half the time and cost of other alternatives.
  • 32
    Rocket Data Virtualization Reviews
    Traditional methods of integrating mainframe, ETL, warehouses, and building connectors are not fast enough or efficient enough to be useful for businesses today. Data virtualization is a new way to store and create more data on the mainframe than ever before. Data virtualization is the only way to close the gap and make mainframe data more accessible to developers and other applications. You can map your data once and then virtualize it for access anywhere, anytime. Your data can scale to your business goals. Data virtualization on z/OS removes the complexity that comes with working with mainframe resources. Data virtualization allows you to combine data from many sources into one logical data source. This makes it easier to connect mainframe data to your distributed applications. Combine mainframe data with location, social networks, and other distributed data.
  • 33
    Informatica PowerCenter Reviews
    The market-leading, scalable, and high-performance enterprise data management platform allows you to embrace agility. All aspects of data integration are supported, from the initial project jumpstart to the successful deployment of mission-critical enterprise applications. PowerCenter, a metadata-driven data management platform, accelerates and jumpstarts data integration projects to deliver data to businesses faster than manual hand coding. Developers and analysts work together to quickly prototype, iterate and validate projects, then deploy them in days instead of months. Your data integration investments can be built on PowerCenter. Machine learning can be used to efficiently monitor and manage PowerCenter deployments across locations and domains.
  • 34
    IBM Cloud Pak for Data Reviews
    Unutilized data is the biggest obstacle to scaling AI-powered decision making. IBM Cloud Pak®, for Data is a unified platform that provides a data fabric to connect, access and move siloed data across multiple clouds or on premises. Automate policy enforcement and discovery to simplify access to data. A modern cloud data warehouse integrates to accelerate insights. All data can be protected with privacy and usage policy enforcement. To gain faster insights, use a modern, high-performance cloud storage data warehouse. Data scientists, analysts, and developers can use a single platform to create, deploy, and manage trusted AI models in any cloud.
  • 35
    erwin Data Catalog Reviews
    Quest's erwin Data Catalog is metadata management software that helps organizations find out what data they have, where it's located, and data in motion and at rest. It provides information about the metadata and data available for a topic, allowing you to quickly find those assets and sources for analysis and decision-making. The erwin Data Catalog automates processes for harvesting, integrating and activating enterprise data in accordance with business requirements. This automation increases accuracy and speeds up time to value for data governance, digital transformation efforts, such as data warehouse, data vault, data lake, and other Big Data deployments, and cloud migrations. Sustainable data governance is essential for any organizational effort that relies on data. erwin Data Catalog automates enterprise data management, data mapping and code generation.
  • 36
    Validio Reviews
    Get a clear view of your data assets: popularity, usage, and schema coverage. Get important insights into your data assets, such as popularity and utilization. Find and filter data based on tags and descriptions in metadata. Get valuable insights about your data assets, such as popularity, usage, quality, and schema cover. Drive data governance and ownership throughout your organization. Stream-lake-warehouse lineage to facilitate data ownership and collaboration. Lineage maps are automatically generated at the field level to help understand the entire data ecosystem. Anomaly detection is based on your data and seasonality patterns. It uses automatic backfilling from historical data. Machine learning thresholds are trained for each data segment and not just metadata.
  • 37
    SAP Data Intelligence Reviews
    Data intelligence can transform data chaos into data value. Disjointed data assets can be connected, discovered, enriched, and orchestrated to provide actionable business insights at an enterprise scale. SAP Data Intelligence provides a comprehensive data management solution. It is the data orchestration layer within SAP's Business Technology Platform. It transforms distributed data into vital data insights and delivers innovation at scale. Integrate across the IT landscape to provide intelligent, relevant and contextual insights for your users. Integrate and orchestrate large data volumes and streams at scale. Machine learning enables you to streamline, operationalize, manage, and govern innovation. Comprehensive metadata management rules optimize governance and reduce compliance risk. Disjointed data assets can be connected, discovered, enriched, and orchestrated to provide actionable business insights at enterprise level.
  • 38
    Enterprise Enabler Reviews
    It unifies information across silos and scattered data for visibility across multiple sources in a single environment; whether in the cloud, spread across siloed databases, on instruments, in Big Data stores, or within various spreadsheets/documents, Enterprise Enabler can integrate all your data so you can make informed business decisions in real-time. By creating logical views from data starting at the source. This allows you to reuse, configure, test and deploy all your data in one integrated environment. You can analyze your business data as it happens to maximize the use and minimize costs, improve/refine business processes, and optimize the use of your assets. Our implementation time to market is between 50-90% shorter. We connect your sources so that you can make business decisions based upon real-time data.
  • 39
    Oracle Big Data Preparation Reviews
    Oracle Big Data Preparation Cloud Service (PaaS), is a cloud-based managed Platform as a Service (PaaS). It allows you to quickly ingest, repair and enrich large data sets in an interactive environment. For down-stream analysis, you can integrate your data to other Oracle Cloud Services such as Oracle Business Intelligence Cloud Service. Oracle Big Data Preparation Cloud Service has important features such as visualizations and profile metrics. Visual access to profile results and summary for each column are available when a data set has been ingested. You also have visual access the duplicate entity analysis results on the entire data set. You can visualize governance tasks on the service homepage with easily understandable runtime metrics, data quality reports and alerts. Track your transforms to ensure that files are being processed correctly. The entire data pipeline is visible, from ingestion through enrichment and publishing.
  • 40
    Oracle Cloud Infrastructure Data Catalog Reviews
    Oracle Cloud Infrastructure (OCI Data Catalog) is a metadata management tool that aids data professionals in discovering data and supporting data governance. It was designed to integrate with the Oracle ecosystem. It provides an inventory of assets and a business glossary. OCI Data Catalog is fully managed and maintained by Oracle. It runs on all the power and scale that Oracle Cloud Infrastructure offers. OCI Data Catalog offers all the security, reliability and performance of Oracle Cloud. Developers can integrate OCI Data Catalog's capabilities into their own applications by using REST APIs or SDKs. Administrators can manage access to OCI Data Catalog objects and security requirements by using a trusted system to manage user identities and access privileges. To get real value out of data, discover data assets in Oracle data stores both on-premises or in the cloud.
  • 41
    Gathr Reviews
    The only platform that can handle all aspects of data pipeline. Gathr was built from the ground up to support a cloud-first world. It is the only platform that can handle all your data integration needs - ingestion and ETL, ELT and CDC, streaming analytics and data preparation, machine-learning, advanced analytics, and more. Gathr makes it easy for anyone to build and deploy pipelines, regardless of their skill level. Ingestion pipelines can be created in minutes and not weeks. You can access data from any source and deliver it to any destination. A wizard-based approach allows you to quickly build applications. A templatized CDC app allows you to replicate data in real time. Native integration for all sources. All the capabilities you need to succeed today or tomorrow. You can choose between pay-per-use, free, or customized according to your needs.
  • 42
    Orbit Analytics Reviews
    A true self-service reporting platform and analytics platform will empower your business. Orbit's business intelligence and operational reporting software is powerful and scalable. Users can create their own reports and analytics. Orbit Reporting + Analytics provides pre-built integration with enterprise resources planning (ERP), key cloud business applications, such as Salesforce, Oracle E-Business Suite and PeopleSoft. Orbit allows you to quickly and efficiently discover answers from any data source, identify opportunities, and make data-driven decisions.
  • 43
    Zaloni Arena Reviews
    End-to-end DataOps built upon an agile platform that protects and improves your data assets. Arena is the leading augmented data management platform. Our active data catalog allows for self-service data enrichment to control complex data environments. You can create custom workflows to increase the reliability and accuracy of each data set. Machine-learning can be used to identify and align master assets for better data decisions. Superior security is assured with complete lineage, including detailed visualizations and masking. Data management is easy with Arena. Arena can catalog your data from any location. Our extensible connections allow for analytics across all your preferred tools. Overcome data sprawl challenges with our software. Our software is designed to drive business and analytics success, while also providing the controls and extensibility required in today's multicloud data complexity.
  • 44
    1touch.io Inventa Reviews
    Your organization is at risk from having only partial visibility to your data. 1touch.io employs a unique network analysis approach, powerful ML/AI techniques, and unparalleled data lineage accuracy to continually discover and catalog all of your sensitive and protected data into a PII inventory and a master data catalog. We automatically detect and analyze all data usage and its lineage without having to rely on the organization's knowledge about the location or existence of the data. Adding a multilayer machine-learning analytic engine allows us to "read and comprehend" the data and link all pieces into a complete picture that can be referred to as both a PII inventory or a Master Catalog. You can reduce your risk by quickly locating sensitive data that is known and unknown within your network. To meet core compliance requirements, you can organize your data flow to understand the exact data lineage and business processes.
  • 45
    Dremio Reviews
    Dremio provides lightning-fast queries as well as a self-service semantic layer directly to your data lake storage. No data moving to proprietary data warehouses, and no cubes, aggregation tables, or extracts. Data architects have flexibility and control, while data consumers have self-service. Apache Arrow and Dremio technologies such as Data Reflections, Columnar Cloud Cache(C3), and Predictive Pipelining combine to make it easy to query your data lake storage. An abstraction layer allows IT to apply security and business meaning while allowing analysts and data scientists access data to explore it and create new virtual datasets. Dremio's semantic layers is an integrated searchable catalog that indexes all your metadata so business users can make sense of your data. The semantic layer is made up of virtual datasets and spaces, which are all searchable and indexed.
  • 46
    erwin Data Intelligence Reviews
    erwin Data Intelligence, or erwin DI, combines data literacy and data catalog capabilities to provide greater awareness and access to data assets, guidance on how to use them, and guardrails to ensure that data policies and best practice are followed. Automatically extract, transform, and feed metadata from a variety of data sources, operational process, and data models into one central catalog. It is then made accessible and understandable through role-based, contextual views. This allows stakeholders to make strategic decisions based upon accurate insights. erwin DI supports enterprise information governance, digital transformation, and any other effort that relies upon data to achieve positive outcomes. You can schedule ongoing scans of metadata from a wide range of data sources. You can easily map data elements from source and target, including data moving, and harmonize data integration across platforms. Data consumers can easily identify and find data that is relevant to their roles.
  • 47
    Oracle Big Data SQL Cloud Service Reviews
    Oracle Big Data SQL Cloud Service allows organizations to instantly analyze data across Apache Hadoop and NoSQL. This service leverages their existing SQL skills, security policy, and applications with extreme speed. Big Data SQL allows you to simplify data science and unlock data lakes. Big Data SQL provides users with a single place to store and secure data in Hadoop, NoSQL systems, and Oracle Database. Seamless metadata integration, and queries that combine data from Oracle Database and Hadoop and NoSQL database data. Automated mappings can be done from metadata stored in HCatalog or the Hive Metastore to Oracle Tables using utility and conversion routines. Administrators have the ability to set enhanced access parameters that allow them to control data access behavior and column mapping. Multiple cluster support allows one Oracle Database to query multiple Hadoop clusters or NoSQL systems.
  • 48
    Hyper-Q Reviews
    Adaptive data virtualization™, a technology that allows enterprises to run existing applications on modern cloud storage warehouses without rewriting them or reconfiguring them, is Adaptive Data Virtualization™. Datometry HyperQ™, a cloud database management software, allows enterprises to adopt new cloud databases quickly, reduce ongoing operating expenses, and develop analytic capabilities to accelerate digital transformation. Datometry HyperQ virtualization software makes it possible to run any existing application on any cloud database. This allows applications and databases to interoperate. Enterprises can now choose the cloud database they prefer, without needing to rip, replace, or rewrite existing applications. Runtime compatibility with legacy data warehouse functions can be achieved through Transformation and Emulation. Transparent deployments on Azure, AWS, or GCP clouds. Applications can continue to use existing JDBC and ODBC connectors. Connects to the major cloud data warehouses Azure Synapse Analytics and AWS Redshift as well as Google BigQuery.
  • 49
    Actifio Reviews
    Integrate with existing toolchain to automate self-service provisioning, refresh enterprise workloads, and integrate with existing tools. Through a rich set APIs and automation, data scientists can achieve high-performance data delivery and re-use. Any cloud data can be recovered at any time, at any scale, and beyond legacy solutions. Reduce the business impact of ransomware and cyber attacks by quickly recovering with immutable backups. Unified platform to protect, secure, keep, govern, and recover your data whether it is on-premises or cloud. Actifio's patented software platform turns data silos into data pipelines. Virtual Data Pipeline (VDP), provides full-stack data management - hybrid, on-premises, or multi-cloud -- from rich application integration, SLA based orchestration, flexible movement, data immutability, security, and SLA-based orchestration.
  • 50
    CData Query Federation Drivers Reviews
    Embedded Data Virtualization allows you to extend your applications with unified data connectivity. CData Query Federation Drivers are a universal data access layer that makes it easier to develop applications and access data. Through a single interface, you can write SQL and access data from 250+ applications and databases. The CData Query Federation Driver provides powerful tools such as: * A Single SQL Language and API: A common SQL interface to work with multiple SaaS and NoSQL, relational, or Big Data sources. * Combined Data Across Resources: Create queries that combine data from multiple sources without the need to perform ETL or any other data movement. * Intelligent Push-Down - Federated queries use intelligent push-down to improve performance and throughput. * 250+ Supported Connections: Plug-and–Play CData Drivers allow connectivity to more than 250 enterprise information sources.