Best Data Management Software for Kedro

Find and compare the best Data Management software for Kedro in 2025

Use the comparison tool below to compare the top Data Management software for Kedro on the market. You can filter results by user reviews, pricing, features, platform, region, support options, integrations, and more.

  • 1
    Jupyter Notebook Reviews
    Open-source web application, the Jupyter Notebook, allows you to create and share documents with live code, equations, and visualizations. Data cleaning and transformation, numerical modeling, statistical modeling and data visualization are just a few of the many uses.
  • 2
    pandas Reviews
    Pandas is an open-source data analysis and manipulation tool that is fast, flexible, flexible, and easy to use. It was built on top the Python programming language. Tools for reading and writing data between memory data structures and various formats: CSV, text files, Microsoft Excel, SQL databases and the fast HDF5 format. Intelligent data alignment and integrated handling missing data: Use a powerful group engine to perform split-apply/combine operations on data sets. Time series-functionality: date range generation and frequency conversion, moving window statistics, date shifting and lagging. You can even create domain-specific offsets and join time sequences without losing data.
  • 3
    Plotly Dash Reviews
    Dash & Dash Enterprise allow you to build and deploy analytic web applications using Python, R, or Julia. No JavaScript or DevOps are required. The world's most successful companies offer AI, ML and Python analytics at a fraction of the cost of full-stack development. Dash is the way they do it. Apps and dashboards that run advanced analytics such as NLP, forecasting and computer vision can be delivered. You can work in Python, R, or Julia. Reduce costs by migrating legacy per-seat license software to Dash Enterprise's unlimited end-user pricing model. You can deploy and update Dash apps faster without an IT or DevOps staff. You can create pixel-perfect web apps and dashboards without having to write any CSS. Kubernetes makes it easy to scale. High availability support for mission-critical Python apps
  • 4
    Vertex AI Reviews
    Fully managed ML tools allow you to build, deploy and scale machine-learning (ML) models quickly, for any use case. Vertex AI Workbench is natively integrated with BigQuery Dataproc and Spark. You can use BigQuery to create and execute machine-learning models in BigQuery by using standard SQL queries and spreadsheets or you can export datasets directly from BigQuery into Vertex AI Workbench to run your models there. Vertex Data Labeling can be used to create highly accurate labels for data collection. Vertex AI Agent Builder empowers developers to design and deploy advanced generative AI applications for enterprise use. It supports both no-code and code-driven development, enabling users to create AI agents through natural language prompts or by integrating with frameworks like LangChain and LlamaIndex.
  • 5
    Matplotlib Reviews
    Matplotlib is a Python library that allows you to create interactive, animated, or static visualizations. Matplotlib makes difficult things simple and easy. Many third-party packages extend and build upon Matplotlib functionality. These include several higher-level plotting interfaces such as seaborn, HoloViews and ggplot.
  • 6
    Apache Spark Reviews

    Apache Spark

    Apache Software Foundation

    Apache Spark™, a unified analytics engine that can handle large-scale data processing, is available. Apache Spark delivers high performance for streaming and batch data. It uses a state of the art DAG scheduler, query optimizer, as well as a physical execution engine. Spark has over 80 high-level operators, making it easy to create parallel apps. You can also use it interactively via the Scala, Python and R SQL shells. Spark powers a number of libraries, including SQL and DataFrames and MLlib for machine-learning, GraphX and Spark Streaming. These libraries can be combined seamlessly in one application. Spark can run on Hadoop, Apache Mesos and Kubernetes. It can also be used standalone or in the cloud. It can access a variety of data sources. Spark can be run in standalone cluster mode on EC2, Hadoop YARN and Mesos. Access data in HDFS and Alluxio.
  • 7
    Dask Reviews
    Dask is free and open-source. It was developed in collaboration with other community projects such as NumPy and pandas. Dask uses existing Python data structures and APIs to make it easy for users to switch between NumPy/pandas and scikit-learn-powered versions. Dask's schedulers can scale to thousands of node clusters, and its algorithms have been tested at some of the most powerful supercomputers around the world. You don't necessarily need a large cluster to get started. Dask ships schedulers that can be used on personal computers. Many people use Dask to scale computations on their laptops, using multiple cores and their disk for extra storage. Dask exposes lower level APIs that allow you to build custom systems for your own applications. This allows open-source leaders to parallelize their own packages, and business leaders to scale custom business logic.
  • 8
    Azure Databricks Reviews
    Azure Databricks allows you to unlock insights from all your data, build artificial intelligence (AI), solutions, and autoscale your Apache Spark™. You can also collaborate on shared projects with other people in an interactive workspace. Azure Databricks supports Python and Scala, R and Java, as well data science frameworks such as TensorFlow, PyTorch and scikit-learn. Azure Databricks offers the latest version of Apache Spark and allows seamless integration with open-source libraries. You can quickly spin up clusters and build in an Apache Spark environment that is fully managed and available worldwide. Clusters can be set up, configured, fine-tuned, and monitored to ensure performance and reliability. To reduce total cost of ownership (TCO), take advantage of autoscaling or auto-termination.
  • 9
    Apache Airflow Reviews

    Apache Airflow

    The Apache Software Foundation

    Airflow is a community-created platform that allows programmatically to schedule, author, and monitor workflows. Airflow is modular in architecture and uses a message queue for managing a large number of workers. Airflow can scale to infinity. Airflow pipelines can be defined in Python to allow for dynamic pipeline generation. This allows you to write code that dynamically creates pipelines. You can easily define your own operators, and extend libraries to suit your environment. Airflow pipelines can be both explicit and lean. The Jinja templating engine is used to create parametrization in the core of Airflow pipelines. No more XML or command-line black-magic! You can use standard Python features to create your workflows. This includes date time formats for scheduling, loops to dynamically generate task tasks, and loops for scheduling. This allows you to be flexible when creating your workflows.
  • Previous
  • You're on page 1
  • Next